Title: Irish Centre for High-End
1Irish Centre for High-End Computing Bringing
capability computing to the Irish research
community
Andy Shearer Director ICHEC
2Why ICHEC?
- Ireland lags well behind the rest of Europe in
terms of installed HEC capacity - See www.arcade-eu.info/academicsupercomputing/comp
arison.html - Ireland now has an installed capacity of about
6000 Gflops/s - Ireland has about15-20 systems above 64 Gflops/s
sustained
3Why ICHEC?
- Ireland - lt2500 processors
- But now about 1500 Kflops/s per capita
4 High-End Computing in Ireland
DFTU spin density. Michael Nolan1, Dean C.
Sayle2, Stephen C. Parker3 and Graeme W.
Watson1(1) TCD, (2) Cranfield University, (3)
University of Bath
Marine Modelling Centre, NUI, Galway
G.Murphy, et al, , DIAS/CosmoGrid
Llyod D., et al, Dept. of Biochemistry, TCD
5Context and Motivation
- Irelands ability to perform internationally
competitive research and to attract the best
computational scientists is currently hindered
due to a lack of high end computational resources - Ireland has the research demands to justify a
large-scale national facility
- The Irish Centre for High-End Computing (ICHEC)
- SFI-funded (2.6M) with in collaboration with the
PRTLI project CosmoGrid (700k), plus links to
the TCD/IITAC cluster. CosmoGrid - Grid-Enabled
Computational Physics of Natural Phenomena. - Initially funded for one year, with proposal to
extend to a minimum of three years
6Participating Institutions
National University of Ireland, Galway Dublin
City University Dublin Institute for Advanced
Studies National University of Ireland,
Maynooth Trinity College Dublin Tyndall
Institute University College Cork University
College Dublin HEAnet
7ICHECs objectives
- Support world-class research programmes within
Ireland - Undertake collaborative research programmes
- Provide access to HEC facilities to students and
researchers in Ireland - Provide HPC training
- Increase the application of HEC and GRID
technologies - Encourage and publicise activities conducive to
establishing and sustaining a world-class HEC
facility in Ireland - Foster collaboration with similar internationally
recognised organisations
8ICHECs Roadmap
Phase I (2005) examine the challenges
associated with the management and operations of
a distributed national HEC centre ? Hire key
personnel, set up infrastructure and policies,
gather user requirements, Phase II (2006)
bring the personnel required in to support such a
system and develop large-scale new science ?
Fully utilise resources, deploy points of
presence, participate in community-led
initiatives, explore prospects for technology
transfer activities, Phase III (2007/8) ICHEC
fully established as a national large-scale
facility
9ICHEC Hardware Resources
- We have a diverse user group so we require
heterogeneous resources. - Our tender looked for
-
- Shared Memory System (20)
- Cluster (60)
- Shared File system (20)
- We were also open to imaginative solutions - as
long as they were within our budget - Acceptable responses from Bull, IBM,
Clustervision, HP, Sun. - Other vendors did not read the tender documents
and/or deliver on time!
and of course if we do this again we would do it
differently - and so I would hope would the
manufacturers
10Shared memory system
- Bull NovaScale NS6320
- 32 Itanium 2 _at_1.5GHz
- 256GB RAM
- 2.1Tb storage
- 192 GFlops/s peak
- 166GFlops/s LINPACK
- Bull Novascale NS4040
- front-end functionalities
- 4 Itanium 2 _at_1.5GHz
11Distributed memory cluster and filesystem
- IBM cluster 1350
- 474 e326 servers
- Dual Opterons _at_2.4GHz each
- 2.2Tb distributed memory
- Mix of (410) 4Gb and (64) 8Gb nodes
- 20Tb storage
- SAN IBM DS4500 using GPFS
- 4.55 TFlops/s peak, 2.8Tflops/s
- Will feature in the upper end of the Top500
- 15 additional nodes
- front-end functionalities and storage/cluster
management - Interconnect initially Gigabit Ethernet
12ICHEC Machine Room
- Installation on schedule
- Acceptance tests mainly completed
- Transitional Service opened 1st September
13 Software
- Establishing user requirements
- discussions with user groups
- Limited software budget
- Prices for National Centres higher than for
academic research groups - NB not possible to purchase all packages on
peoples wish list - We had pressure to purchase many packages with
overlapping functionalities need to rationalise - Preference given to packages with best
scalability/performance - Will work with research communities to reduce
this list
14 Software (cont.)
- Currently shortlisted 30 packages
- Current list
- Operating systems
- SUSE Linux Enterprise 9, Bull Linux (based on Red
Hat Enterprise) - Programming environments
- Intel compilers, Portland Group Cluster
Development Kit, DDT - Pre-/Post-processing, data visualisation
- IDL, Matlab, OpenDX, VTK
- Libraries and utilities
- FFTW, Scalapack, PetSC, (Par)Metis
15 Software (cont.)
- Engineering software
- CFX, Abaqus, Marc, Patran
- Computational chemistry / materials sciences /
life sciences - Amber, (MPI-)BLAST, Charmm, CPMD, Crystal,
DLPoly, Gamess UK, Gaussian03, Gromacs, MOE,
Molpro, NAMD, NWchem, Siesta, Turbomole, VASP - Environmental sciences
- POM
- Expensive packages (Accelerys, etc.) not
initially available - defer decision until Phase 2 budget approved
16 Scheduling Policies
- Transitional Period, cluster
- Day regime week days, 10am 6pm
- Night regime week days, 6pm 10am
- Week-end regime Friday 6pm Monday 10am
- Tuning pending
- Scalability studies of key applications
- User feedback
- Gaussian users expected to use the Production
region / HiMem - Allow acceptable turn-around time on the Bull for
FEM users
17 Scheduling Policies
- Transitional Period, shared-memory system
- Day/night/week-end regimes same as cluster
- Feedback from user community is important to help
us tune policies for the start of the full
National Service - Need for checkpoint/re-start capability for long
runs - Fair share mechanism enforced by Maui
- Mechanism to jump the queue for important
deadlines
18Supporting the research community
- ICHEC will be set up as a distributed national
centre - The Centre will be an equally important mix of
Hardware and People - Large groups in Dublin, Galway and Cork
- Points of presence in each of the participating
institutions (Phase 2) - Support will be essential
- Plans are to recruit applications scientists in
scientific areas - engineering (FEM/CFD/LB)
- life sciences (bioinformatics, drug design)
- physics /chemistry (soft and hard condensed
matter) - environmental sciences (geology/geophysics,
oceanography, meteorology/climatology) - applied maths
- HPC
19Supporting the research community (cont.)
- The face of the Centre will be the support
scientists - Serve their research community, rather than
providing free staff to individual groups - Code development (community-led initiative)
- Make the best use of the infrastructure
(parallelise/optimise) - Development of specialised material (course,
documentation) - Technical support for Grand Challenge projects
- Persistent communication channel with ICHEC
- Key to our success will be a close and sustained
collaboration with the research community
20Training activities
- Training is seen as a way of increasing
performance and efficiency of the machines - our
programme reflects the relative youth of
Irelands HEC activity. - ICHEC has developed courses - the first was
delivered to UCD in October. - An introduction to High-Performance Computing
(HPC) - What is HPC, HPC as a tool for furthering
research, etc. - Overview of current hardware architectures
- Overview of common programming paradigms
- Decomposition strategies
- An introduction to the ICHEC national service
- Parallel programming with MPI an introduction
- Parallel programming with OpenMP an introduction
- Specialised material to be developed in 2006,
e.g. - HPC for engineers, HPC in bioinformatics,
parallel linear algebra, etc.
21Now
- Service started on the 1st September (Bull)
- 21st September (IBM)
- Have 60 approved projects plus 17 from CosmoGrid
- 38 Physical Sciences
- 22 Engineering
- 25 Life Sciences
- 7 Environmental Sciences
- 7 Other (Maths, Computer Sciences, Humanities)
- At least three times as many applications will be
submitted in 3-6 months - Running at full capacity by the end of the year
22Problems
- Getting the community up to speed on
- Batch submission
- Moving from serial to parallel computing
- Licences?
- National software agreements
- Firewalls and security
- Having a miriad of security and firewall
standards for HEAnet is insanity - DCU and TCD do not accept zip files but have
different ways around this - Proxy servers for ssh is different on each site
- Grid access is a nightmare .
- A common approach via HEAnet would help
- The last mile problem
- Our access is fine, redundant Gigabit but the
universities throttle this back
23The future enhanced hardware
- Technology/Architecture?
- Massive clusters bring their own problems of
reliability - Large scale SMP vs tightly bound clusters of
fat nodes - Novel architecture - IBMs BlueGene or FPGAs -
others - Software?
- Massive clusters bring massive problems
- Scalability
- Fault tolerance - how to do a check point restart
on 10000 nodes? - Licences?
- How best to licence software on a 10000 node
cluster? - How to benchmark?
- In 2006/7 we will be looking for original
solutions to these problems