SlideShare a Scribd company logo
1 of 35
Download to read offline
40 powers of 10
Simulating the Universe from
quarks to galaxy clusters with the
DiRAC HPC Facility
Dr Mark Wilkinson
University of Leicester
Director, DiRAC HPC Facility
DiRAC
DiRAC
Distributed HPC Facility for theoretical astrophysics,
particle physics, cosmology and nuclear physics
Combined:
• ~90,000 cores
• ~5 Pflop/s
• >10 PB storage
Extreme Scaling (Edinburgh)
Data Intensive (Cambridge)
Data Intensive
(Leicester)
Memory Intensive (Durham)
The DiRAC Facility, a brief history
2009: DiRAC-1: systems installed at 13 host sites
Nov 2011: DiRAC-2 awarded £15M capital by BIS
- 5 systems at 4 sites - Cambridge, Durham, Edinburgh, Leicester
- Procurement completed in 100 days
Dec 2012: DiRAC-2 operations begin
- Systems full within 1 week - usage >90%
- Access via international peer-review process
- Free to use for STFC-funded researchers
April 2017: DiRAC-2.5 operations begin
- 3 services: Extreme Scaling, Memory Intensive, Data Intensive
April 2018: DiRAC-2.5x
- interim funding to replace 2012 hardware and support 2018/19
science programme
Dec 2018: DiRAC-2.5y
- BEIS-funded stop-gap 2x upgrade of all DiRAC services
• Direct numerical simulation and modelling is core theory research activity
• HPC systems are main scientific instruments for theory research
• Computational requirements of models increasing due to
• Increased resolution: running models with existing physics at finer
scales.
• Increased complexity: introducing new physics into models to reflect
progress in theoretical understanding; often needed to match resolution
• Coupling of models: multi-physics, multi-scale modelling;
• Quantification of modelling uncertainty using large ensembles of
simulations to provide robust statistics
• Constant process of refining and re-defining our tools
• Growing requirement for simulations and modelling concurrent with
observations so that models evolve in line with data acquired
• Observational facilities need access to significant local computing
capabilities as well as option to burst out to the larger, national facilities.
Computing in theory research
The DiRAC approach to service design
Science
case
Workflow
assessment
Technical
case
Peer-reviewed, scientific justification for resources
Peer-reviewed, high-level technical specifications
Technical
design
Co-design with
industry partners
Individual service specifications
Diverse science cases require
heterogenous architectures
Extreme Scaling
“Tesseract”
(Edinburgh)
Memory Intensive
“COSMA”
(Durham)
Data Intensive
“DIaL” and “CSD3”
(Leicester & Cambridge)
230 TB RAM to
support largest
cosmological
simulations
Heterogeneous
architecture to support
complex simulation and
modelling workflows
2 Pflop/s to
support largest
lattice-QCD
simulations
DiRAC
scientific impact is highly significant, sustained and of world leading quality.
Figure 3-7: Facility Citations and Paper Record 2015-2018
3.5 WP5: Project Management
DiRAC Science Impact Analysis
• Refereed publications since 2012: > 950 papers; > 49000 citations
• Second-most cited paper in all astronomy in 2015
• 199 refereed papers in 2017
DiRAC@Edinburgh – Extreme Scaling (ES)
• HPE SGI 8600
• >1.8 PF
• 1468 standard compute nodes
• Two 2.1GHz 12-core Intel Xeon Skylake 4116 processors
• 96GB RAM
• 8 GPU compute nodes
• Two 2.1GHz 12-core Intel Xeon Skylake 4116 processors
• 96 GB RAM
• 4 NVidia V100 (Volta) GPU accelerators connected over PCIe
• Hypercube Intel Omnipath interconnect
• 3PB Lustre DDN storage
“Tesseract”
The forces of nature
Inside the proton
Image: Arpad Horvath - Own work, CC BY-SA 2.5,
https://commons.wikimedia.org/w/index.php?curid=637353
quark
Strong interactions
carried by gluons
Understanding the strong force is key to testing the
Standard Modal of Particle Physics - binds quarks into
hadrons that we see in experiment
ATLAS@Large Hadron
Collider (LHC)
Connecting observed hadron
properties to those of quarks
requires full non-perturbative
treatment of Quantum
Chromodynamics - lattice QCD
DALI Run=16449 Evt=4055
PH
1cm|
IP
B
D
ππ
KK
K
π
e−
−
−
+
+
+s
s
Calculations of
hadron masses and
rates for simple weak
decays allow tests of
Standard Model
• L4
local volume (space + time)
• finite di↵erence operator 8 point stencil
• ⇠ 1
L of data references come from o↵ node
Scaling QCD sparse matrix requires interconnect bandwidth for halo exchange
Bnetwork ⇠
Bmemory
L
⇥R
where R is the reuse factor obtained for the stencil in caches
• Aim: Distribute 1004
datapoints over 104
nodes
Scalable QCD: sparse matrix PDE solver communications (Boyle et al.)
Immediate Big HPC roadmap
0.1
1
10
100
1000
2010 2012 2014 2016 2018 2020 2022
PF/system
IBM-BGQ/Mira
Intel-Broadwell/Cori1
Intel-KNL/Cori2
Intel-KNL/Oakforest PACS
IBM/Nvidia/Summit/Sierra
Intel-KNH/Aurora
Fujitsu-ARM/post-K
Frontier/ORNL
0.1
1
10
100
1000
2010 2012 2014 2016 2018 2020 2022
SPTF/node
IBM-BGQ/Mira
Intel-Broadwell/Cori1
Intel-KNL/Cori2
Intel-KNL/Oakforest PACS
IBM/Nvidia/Summit
0.1
1
10
100
1000
2010 2012 2014 2016 2018 2020 2022
commsGB/s/node
IBM-BGQ/Mira
Intel-Broadwell/Cori1
Intel-KNL/Cori2
Intel-KNL/Oakforest PACS
IBM/Nvidia/Summit/Sierra
• 400x+ increase in SP node performance accompanied by NO increase in interconnect
Immediate “Big HPC” roadmap
Grid QCD code
Design considerations
• Performance portable across multi and many core CPU’s
SIMD⌦OpenMP⌦MPI
• Performance portable to GPU’s
SIMT⌦offload⌦MPI
• N-dimensional cartesian arrays
• Multiple grids
• Data parallel C++ layer : Connection Machine inspired
Grid QCD code
(Boyle et al.)
• Relatively cheap node: high node count and scalability
• Improve price per core, bandwidth per core, reduce power
Tesseract performance per node vs nodes, volume
GF/spernode
0.0
150.0
300.0
450.0
600.0
Nodes
1 16 32 64 128 256
12^4 16^4 24^4
nodes (single switch) delivers bidirectional 25GB/s to every node (wirespeed)
2 nodes topology aware bidirectional 19GB/s
• 76% wirespeed using every link in system concurrently
DiRAC HPE ICE-XA hypercube network
• Edinburgh HPE 8600 system (Installed March 2018, 2 days ahead of schedule)
• Low end Skylake Silver 4116, 12 core parts
• Single rail Omnipath interconnect
• Relatively cheap node: high node count and scalability
• Improve price per core, bandwidth per core, reduce power
Tesseract performance per node vs nodes, volume
GF/spernode
0.0
150.0
300.0
450.0
600.0
Nodes
1 16 32 64 128 256
12^4 16^4 24^4
• 16 nodes (single switch) delivers bidirectional 25GB/s to every node (wirespeed)
• 512 nodes topology aware bidirectional 19GB/s
• 76% wirespeed using every link in system concurrently
• Low end Skylake Silver 4116, 12 core parts
• Single rail Omnipath interconnect
• Relatively cheap node: high node count and scalability
• Improve price per core, bandwidth per core, reduce power
Tesseract performance per node vs nodes, volume
GF/spernode
0.0
150.0
300.0
450.0
600.0
Nodes
1 16 32 64 128 256
12^4 16^4 24^4
• 16 nodes (single switch) delivers bidirectional 25GB/s to every node (wirespeed)
• 512 nodes topology aware bidirectional 19GB/s
• 76% wirespeed using every link in system concurrently
• Low end Skylake Silver 4116, 12 core parts
• Single rail Omnipath interconnect
• Relatively cheap node: high node count and scalability
• Improve price per core, bandwidth per core, reduce power
Tesseract performance per node vs nodes, volume
GF/spernode
0.0
150.0
300.0
450.0
600.0
Nodes
1 16 32 64 128 256
12^4 16^4 24^4
16 nodes (single switch) delivers bidirectional 25GB/s to every node (wirespeed)
512 nodes topology aware bidirectional 19GB/s
• 76% wirespeed using every link in system concurrently
• Small project with SGI/HPE on Mellanox EDR networks (James Southern)
• Embed 2n
QCD torus inside hypercube so that nearest neigbour comms travels single hop
4x speed up over default MPI Cartesian communicators on large systems
) Customise HPE 8600 (SGI ICE-XA) to use 16 = 24
nodes per leaf switch
Boyle et al.
Other Turing work
• Intel’s US Patent Application 20190042544
(FP16-S7E8 MIXED PRECISION FOR DEEP LEARNING AND OTHER ALGORITHMS)
• Authors: Boyle (ATI, Edinburgh), Sid Kashyap, Angus Lepper (Intel, former DiRAC RSE’s)
• Systematic study using Gnu Multi Precision library.
• BFP16 displays greater numerical stability for machine learning training.
• Understanding: histogramme of multiply results during SGD gradient calculations
• Patent application full text searchable on uspto.gov
DiRAC / Alan Turing Institute / Intel Collaboration
Software Innovation - AI on HPC
cessors.
������
�����
����
���
�
�� ��� ���� ����� ������ ����� ����� �����
�������
�����
����� ���� �����
����� ���� �������
G. 5: Wall clock time per reduction call vs. vector length before and after our optimisation. The large vector reducti
formance is ten times better after our optimisations on large vector lengths. The gain includes both computation accelerati
d communication acceleration.
the interface to memory management is not ideal.
ther, it is likely that method dealloc and alloc routine provided by collectives.h should have been used consistent
th use of any other “free” operation declared illegal. Further, this would have enabled simpler modification of t
ocation and deallocation implementation. It would also have been better to separate the vector allocation from t
duction operation; so that in hot loops the programmer could choose to reuse the same allocation.
ur first optimisations were high level: i) remove the expectation that the caller deallocates the returned vector;
Boyle et al.
2017
• Demonstration of factor 10 speed-up in the Baidu Research
optimised reduction code 

- a publicly available code designed to optimise the performance
limiting steps in distributed machine learning (ML)
• potentially disruptive implications for design of cloud systems 

- shows that ML workflows can achieve 10x performance
improvement when fully optimised and implemented on traditional
HPC architectures
gure 1: Left panel: A ‘fan’ plot for the light quark ud and us decay constants (normalised with the
erage ud, us decay constant) for one lattice spacing against the pion mass; Right panel: The heavy b
uark ub, sb decay constants, again normalised, for several lattice spacings also plotted against the
on mass.
ow further extended this programme to the heavier quark masses and in particular the b, [2].
the right panel of the figure we give the equivalent plot for the B and Bs mesons, together
ith the present (FLAG16) values.
Bornyakov et al. 2017 Hollitt et al. 2018
Testing the standard model of particle physics
• Mesons are composed of quark-antiquark pairs

• Their decay rates into other particles provide a strong test of the standard model

• Anomalies may point to new physics

• DiRAC researchers have developed new, computationally efficient technique for
estimating decay rates
DiRAC@Durham - Memory Intensive (MI)
• COSMA6 (2016)
• IBM/Lenovo/DDN
• 8192 Intel SandyBridge, 2.6 GHz cores
• Combined 65 Tbyte of RAM;
• Mellanox FDR10 interconnect in 2:1 blocking;
• 2.5 Pbyte of Lustre data storage
• COSMA7 (2018)
• Dell
• 23,600 Intel Skylake 4116 cores
• Interconnect: Mellanox EDR in a 2:1 blocking configuration with
islands of 24 nodes;
• Combined 230 Tbyte of RAM;
• a fast checkpointing i/o system (343 Tbyte) with peak performance
of 185 Gbyte/second write and read;
• >4 PB storage
Durham – MI
m was delivered by Dell in March 2018;
y Alces
service started on 1 May 2018
engagement:
closely with the Industrial Strategy of
partment for Business, Energy and
ial Strategy (BEIS).
g leads to industrial engagement; this
in innovation, both leading to
ts academia and the wider industry.
The Universe, a brief history
Credit: NASA / WMAP Science Team
DiRAC supports
calculations on all
physical scales
The components of the Universe
Credit: NASA / WMAP Science Team & Planck Satellite Team
26.8%
4.9%
68.3%
ICC Durham; Virgo Consortium
The EAGLE Simulation
DiRAC
Dark matter (N-body)
The EAGLE Simulation
ICC Durham; Virgo ConsortiumDiRAC
Gas and stars (hydrodynamics)
Research Software Engineers
“With great power comes great responsibility”
• Science requirements for DiRAC-3 demand 10-40x increases in
computing power to stay competitive
- hardware alone cannot deliver this
• We can no longer rely on “free lunch” from the Xeon era
• Vectorisation and code efficiency now critical
• Next generation hardware is more difficult to program efficiently
• RSEs are increasingly important
- RSEs can help with code profiling, optimisation, porting, etc
• DiRAC now has 3 RSEs: effort allocated via peer review process
algorithmic
computational
David Keyes
FirstGalaxies
SolarSystemForms
Reionization
Milky Way
13121110987654321
Age of the Universe (Gyr)
Present DayBig Bang
The Evolution of the Universe
First simulation on DiRAC-2.5y Memory
Intensive system at Durham carried out
with SWIFT, a revolutionary new
cosmological hydrodynamics code
developed at Durham which is 20x faster
than the state-of-the-art.
Key Problems
Origin of galaxies
Indentity of the dark matter
Nature of the dark energy
• A fast checkpointing i/o system (343 Tbyte) with peak performance of
185 Gbyte/second write and read
• 15 Lustre Object Storage Servers on Dell 640 nodes:
• 2 x Intel Skylake 5120 processors
• 192 Gbyte of RAM
• 8 x 3.2 TB NVMe SFF drives.
• 1 Mellanox EDR card
• A user code benchmark produced 180 Gbyte/second write and read –
this is almost wire-speed!
• At time of installation this was the fastest filesystem in production in
Europe
Memory Intensive:
burst buffer for checkpointing
0
500,000
1,000,000
1,500,000
2,000,000
2,500,000
3,000,000
24-hour 12-hour 6-hour 4-hour 2-hour 1-hour
KiloWatthoursover5years
Axis Title
Power usage for snapshots with different performance solutions
30GB/sec 120GB/sec 140GB/sec
Memory Intensive:
burst buffer for checkpointing
Power saving
Heck (2018)
Core hours lost in checkpointing over 5 years
Snapshot
period 24-hr 12-hr 6-hr 4-hr 2-hr 1-hr
Hours/
snapshot
30GB/sec 7.1M 14.2M 28.4M 42.5M 85.1M 170.1M 0.95
120GB/sec 1.8M 3.5M 7.1M 10.6M 21.3M 42.5M 0.24
140GB/sec 1.5M 3.0M 6.1M 9.1M 18.2M 36.54M 0.20
• Total number of available cpu hours per year: 36M (4116 cores) 
• ~13% gain in available core hours due to faster checkpointing
Memory Intensive:
burst buffer for checkpointing
Heck (2018)
Uranus collision?
8 J. A. Kegerreis et al.
10 5 0
x Position (R )
10
5
0
yPosition(R)
10 5 0
x Position (R )
106
107
108
SpecificInternalEnergyJkg1
Figure 6. A mid-collision snapshots of a grazing impact with 108 SPH particles – compared with the more head-on collision in Fig. 5 –
coloured by their material and internal energy, showing some of the detailed evolution and mixing that can now be resolved. In the left
panel, light and dark grey show the target’s ice and rock material, respectively, and purple and brown show the same for the impactor.
Light blue is the target’s H-He atmosphere.
mon aim in cosmology, to simulate a larger patch of the uni-
verse), or we can study a small system with higher resolution
to model smaller details.
not) question that we ask a computer to solve, so this is an
important first step.
Kegerreis et al.
(2019), subm.
Planet collisions
Giant impact explains Uranus spin axis and cold temperature
• 108-particle simulations carried out on DiRAC Memory Intensive service
(Durham) using new hydro+Nbody code, Swift
• Hardware and software for this calculation developed with DiRAC support.
DiRAC & the discovery of Gravitational waves
On September 14, 2015 at 09:50:45 UTC, the LIGO
Hanford, WA, and Livingston, LA, observatories detected
pact binary waveforms [44] recovered GW
most significant event from each detector f
tions reported here. Occurring within the
• Simulations of binary
black hole mergers
performed on DiRAC
DataCentric system
(COSMA5)
• Crucial for interpretation
of LIGO gravitational
wave detection
Abbott et al., (2016)
DiRAC
DiRAC@Leicester – Data Intensive (DIaL)
• HPE system
• 400 compute nodes
• 2x Intel Xeon Skylake 6140, 2.3GHz, 18-core processors
• Dual FMA AVX512
• 192 GB RAM
• 1 x 6TB SuperDome Flex with 144 X6154 cores (3.0GHz)
• 3 x 1.5TB fat nodes with 36 X6140 cores
• Mellanox EDR interconnect in a 2:1 blocking setup
• 3PB Lustre storage.
• 150 TB flash storage for data intensive workflows
HPE/Arm/Suse Catalyst UK Arm system
• 4000-core Thunder X2-based cluster installed in January 2019
• Infiniband all-to-all interconnect
Shared platform with local Cambridge and EPSRC Tier 2 clusters
• Dell multi-architecture system (Skylake, KNL, Nvidia GPU)
• 484 Skylake nodes
• 2 x Intel Xeon Skylake 6142 2.6GHz 16-core processors
• Mix of 192 GB/node and 384 GB/node
• 44 Intel KNL nodes
• Intel Xeon Phi CPU 7210 @ 1.30GHz
• 96GB of RAM per node
• 12 NVIDIA GPU nodes
• four NVIDIA Tesla P100 GPUs
• 96GB memory per node
• connected by Mellanox EDR Infiniband
• Intel OmniPath interconnect in 2:1 blocking configuration
• 1.5PB of Lustre disk storage;
• 0.5PB flash storage-based “Data Accelerator”
DiRAC@Cambridge – Data Intensive (DIaC)
Star	Cluster	Forma.on	in	Clouds	with	Different	Metallici.es	(Bate	2019):		Gas	Temperatures
DiRAC HPC Training
• DiRAC provides access to training from wide pool of providers
• Currently offering:
- DiRAC Essentials Test: now available online (and
compulsory!)
- Workshops and Hackathons
• Coming soon:
- Domain-specific workshops
- Online individual training portal
Why do we do this?
- maximise DiRAC science output
- flexibility to adopt most cost-effective technologies
- future-proofing our software and skills
- contributes to increasing skills of wider economy
(Better systems) + (Better software) = Better science
This requires:
• More engagement in hardware and software co-design
• Enhanced training and knowledge transfer
DiRAC

More Related Content

What's hot

Academic cloud experiences cern v4
Academic cloud experiences cern v4Academic cloud experiences cern v4
Academic cloud experiences cern v4Tim Bell
 
Massively Parallel K-Nearest Neighbor Computation on Distributed Architectures
Massively Parallel K-Nearest Neighbor Computation on Distributed Architectures Massively Parallel K-Nearest Neighbor Computation on Distributed Architectures
Massively Parallel K-Nearest Neighbor Computation on Distributed Architectures Intel® Software
 
How to Design Scalable HPC, Deep Learning, and Cloud Middleware for Exascale ...
How to Design Scalable HPC, Deep Learning, and Cloud Middleware for Exascale ...How to Design Scalable HPC, Deep Learning, and Cloud Middleware for Exascale ...
How to Design Scalable HPC, Deep Learning, and Cloud Middleware for Exascale ...inside-BigData.com
 
TeraGrid and Physics Research
TeraGrid and Physics ResearchTeraGrid and Physics Research
TeraGrid and Physics Researchshandra_psc
 
Hpc, grid and cloud computing - the past, present, and future challenge
Hpc, grid and cloud computing - the past, present, and future challengeHpc, grid and cloud computing - the past, present, and future challenge
Hpc, grid and cloud computing - the past, present, and future challengeJason Shih
 
High Performance Computing - Challenges on the Road to Exascale Computing
High Performance Computing - Challenges on the Road to Exascale ComputingHigh Performance Computing - Challenges on the Road to Exascale Computing
High Performance Computing - Challenges on the Road to Exascale ComputingHeiko Joerg Schick
 
Experiences in Application Specific Supercomputer Design - Reasons, Challenge...
Experiences in Application Specific Supercomputer Design - Reasons, Challenge...Experiences in Application Specific Supercomputer Design - Reasons, Challenge...
Experiences in Application Specific Supercomputer Design - Reasons, Challenge...Heiko Joerg Schick
 
Petascale Analytics - The World of Big Data Requires Big Analytics
Petascale Analytics - The World of Big Data Requires Big AnalyticsPetascale Analytics - The World of Big Data Requires Big Analytics
Petascale Analytics - The World of Big Data Requires Big AnalyticsHeiko Joerg Schick
 
Dynamic Resource Allocation Algorithm using Containers
Dynamic Resource Allocation Algorithm using ContainersDynamic Resource Allocation Algorithm using Containers
Dynamic Resource Allocation Algorithm using ContainersIRJET Journal
 
CloudLightning and the OPM-based Use Case
CloudLightning and the OPM-based Use CaseCloudLightning and the OPM-based Use Case
CloudLightning and the OPM-based Use CaseCloudLightning
 
MVAPICH: How a Bunch of Buckeyes Crack Tough Nuts
MVAPICH: How a Bunch of Buckeyes Crack Tough NutsMVAPICH: How a Bunch of Buckeyes Crack Tough Nuts
MVAPICH: How a Bunch of Buckeyes Crack Tough Nutsinside-BigData.com
 
Science DMZ
Science DMZScience DMZ
Science DMZJisc
 
Towards a Lightweight Multi-Cloud DSL for Elastic and Transferable Cloud-nati...
Towards a Lightweight Multi-Cloud DSL for Elastic and Transferable Cloud-nati...Towards a Lightweight Multi-Cloud DSL for Elastic and Transferable Cloud-nati...
Towards a Lightweight Multi-Cloud DSL for Elastic and Transferable Cloud-nati...Nane Kratzke
 
Scalable Deep Learning in ExtremeEarth-phiweek19
Scalable Deep Learning in ExtremeEarth-phiweek19Scalable Deep Learning in ExtremeEarth-phiweek19
Scalable Deep Learning in ExtremeEarth-phiweek19ExtremeEarth
 
Blue Waters and Resource Management - Now and in the Future
 Blue Waters and Resource Management - Now and in the Future Blue Waters and Resource Management - Now and in the Future
Blue Waters and Resource Management - Now and in the Futureinside-BigData.com
 
Benchmark Analysis of Multi-core Processor Memory Contention April 2009
Benchmark Analysis of Multi-core Processor Memory Contention April 2009Benchmark Analysis of Multi-core Processor Memory Contention April 2009
Benchmark Analysis of Multi-core Processor Memory Contention April 2009James McGalliard
 
Lug best practice_hpc_workflow
Lug best practice_hpc_workflowLug best practice_hpc_workflow
Lug best practice_hpc_workflowrjmurphyslideshare
 
産総研におけるプライベートクラウドへの取り組み
産総研におけるプライベートクラウドへの取り組み産総研におけるプライベートクラウドへの取り組み
産総研におけるプライベートクラウドへの取り組みRyousei Takano
 
DATE 2020: Design, Automation and Test in Europe Conference
DATE 2020: Design, Automation and Test in Europe ConferenceDATE 2020: Design, Automation and Test in Europe Conference
DATE 2020: Design, Automation and Test in Europe ConferenceLEGATO project
 

What's hot (20)

ECP Application Development
ECP Application DevelopmentECP Application Development
ECP Application Development
 
Academic cloud experiences cern v4
Academic cloud experiences cern v4Academic cloud experiences cern v4
Academic cloud experiences cern v4
 
Massively Parallel K-Nearest Neighbor Computation on Distributed Architectures
Massively Parallel K-Nearest Neighbor Computation on Distributed Architectures Massively Parallel K-Nearest Neighbor Computation on Distributed Architectures
Massively Parallel K-Nearest Neighbor Computation on Distributed Architectures
 
How to Design Scalable HPC, Deep Learning, and Cloud Middleware for Exascale ...
How to Design Scalable HPC, Deep Learning, and Cloud Middleware for Exascale ...How to Design Scalable HPC, Deep Learning, and Cloud Middleware for Exascale ...
How to Design Scalable HPC, Deep Learning, and Cloud Middleware for Exascale ...
 
TeraGrid and Physics Research
TeraGrid and Physics ResearchTeraGrid and Physics Research
TeraGrid and Physics Research
 
Hpc, grid and cloud computing - the past, present, and future challenge
Hpc, grid and cloud computing - the past, present, and future challengeHpc, grid and cloud computing - the past, present, and future challenge
Hpc, grid and cloud computing - the past, present, and future challenge
 
High Performance Computing - Challenges on the Road to Exascale Computing
High Performance Computing - Challenges on the Road to Exascale ComputingHigh Performance Computing - Challenges on the Road to Exascale Computing
High Performance Computing - Challenges on the Road to Exascale Computing
 
Experiences in Application Specific Supercomputer Design - Reasons, Challenge...
Experiences in Application Specific Supercomputer Design - Reasons, Challenge...Experiences in Application Specific Supercomputer Design - Reasons, Challenge...
Experiences in Application Specific Supercomputer Design - Reasons, Challenge...
 
Petascale Analytics - The World of Big Data Requires Big Analytics
Petascale Analytics - The World of Big Data Requires Big AnalyticsPetascale Analytics - The World of Big Data Requires Big Analytics
Petascale Analytics - The World of Big Data Requires Big Analytics
 
Dynamic Resource Allocation Algorithm using Containers
Dynamic Resource Allocation Algorithm using ContainersDynamic Resource Allocation Algorithm using Containers
Dynamic Resource Allocation Algorithm using Containers
 
CloudLightning and the OPM-based Use Case
CloudLightning and the OPM-based Use CaseCloudLightning and the OPM-based Use Case
CloudLightning and the OPM-based Use Case
 
MVAPICH: How a Bunch of Buckeyes Crack Tough Nuts
MVAPICH: How a Bunch of Buckeyes Crack Tough NutsMVAPICH: How a Bunch of Buckeyes Crack Tough Nuts
MVAPICH: How a Bunch of Buckeyes Crack Tough Nuts
 
Science DMZ
Science DMZScience DMZ
Science DMZ
 
Towards a Lightweight Multi-Cloud DSL for Elastic and Transferable Cloud-nati...
Towards a Lightweight Multi-Cloud DSL for Elastic and Transferable Cloud-nati...Towards a Lightweight Multi-Cloud DSL for Elastic and Transferable Cloud-nati...
Towards a Lightweight Multi-Cloud DSL for Elastic and Transferable Cloud-nati...
 
Scalable Deep Learning in ExtremeEarth-phiweek19
Scalable Deep Learning in ExtremeEarth-phiweek19Scalable Deep Learning in ExtremeEarth-phiweek19
Scalable Deep Learning in ExtremeEarth-phiweek19
 
Blue Waters and Resource Management - Now and in the Future
 Blue Waters and Resource Management - Now and in the Future Blue Waters and Resource Management - Now and in the Future
Blue Waters and Resource Management - Now and in the Future
 
Benchmark Analysis of Multi-core Processor Memory Contention April 2009
Benchmark Analysis of Multi-core Processor Memory Contention April 2009Benchmark Analysis of Multi-core Processor Memory Contention April 2009
Benchmark Analysis of Multi-core Processor Memory Contention April 2009
 
Lug best practice_hpc_workflow
Lug best practice_hpc_workflowLug best practice_hpc_workflow
Lug best practice_hpc_workflow
 
産総研におけるプライベートクラウドへの取り組み
産総研におけるプライベートクラウドへの取り組み産総研におけるプライベートクラウドへの取り組み
産総研におけるプライベートクラウドへの取り組み
 
DATE 2020: Design, Automation and Test in Europe Conference
DATE 2020: Design, Automation and Test in Europe ConferenceDATE 2020: Design, Automation and Test in Europe Conference
DATE 2020: Design, Automation and Test in Europe Conference
 

Similar to Simulating the Universe from quarks to galaxy clusters with DiRAC HPC

TeraGrid Communication and Computation
TeraGrid Communication and ComputationTeraGrid Communication and Computation
TeraGrid Communication and ComputationTal Lavian Ph.D.
 
Dcn invited ecoc2018_short
Dcn invited ecoc2018_shortDcn invited ecoc2018_short
Dcn invited ecoc2018_shortShuangyi Yan
 
Accelerating TensorFlow with RDMA for high-performance deep learning
Accelerating TensorFlow with RDMA for high-performance deep learningAccelerating TensorFlow with RDMA for high-performance deep learning
Accelerating TensorFlow with RDMA for high-performance deep learningDataWorks Summit
 
Panel: NRP Science Impacts​
Panel: NRP Science Impacts​Panel: NRP Science Impacts​
Panel: NRP Science Impacts​Larry Smarr
 
The Coming Age of Extreme Heterogeneity in HPC
The Coming Age of Extreme Heterogeneity in HPCThe Coming Age of Extreme Heterogeneity in HPC
The Coming Age of Extreme Heterogeneity in HPCinside-BigData.com
 
The SKA Project - The World's Largest Streaming Data Processor
The SKA Project - The World's Largest Streaming Data ProcessorThe SKA Project - The World's Largest Streaming Data Processor
The SKA Project - The World's Largest Streaming Data Processorinside-BigData.com
 
Barcelona Supercomputing Center, Generador de Riqueza
Barcelona Supercomputing Center, Generador de RiquezaBarcelona Supercomputing Center, Generador de Riqueza
Barcelona Supercomputing Center, Generador de RiquezaFacultad de Informática UCM
 
Linac Coherent Light Source (LCLS) Data Transfer Requirements
Linac Coherent Light Source (LCLS) Data Transfer RequirementsLinac Coherent Light Source (LCLS) Data Transfer Requirements
Linac Coherent Light Source (LCLS) Data Transfer Requirementsinside-BigData.com
 
Design installation-commissioning-red raider-cluster-ttu
Design installation-commissioning-red raider-cluster-ttuDesign installation-commissioning-red raider-cluster-ttu
Design installation-commissioning-red raider-cluster-ttuAlan Sill
 
DARPA ERI Summit 2018: The End of Moore’s Law & Faster General Purpose Comput...
DARPA ERI Summit 2018: The End of Moore’s Law & Faster General Purpose Comput...DARPA ERI Summit 2018: The End of Moore’s Law & Faster General Purpose Comput...
DARPA ERI Summit 2018: The End of Moore’s Law & Faster General Purpose Comput...zionsaint
 
Java Thread and Process Performance for Parallel Machine Learning on Multicor...
Java Thread and Process Performance for Parallel Machine Learning on Multicor...Java Thread and Process Performance for Parallel Machine Learning on Multicor...
Java Thread and Process Performance for Parallel Machine Learning on Multicor...Saliya Ekanayake
 
Exploring emerging technologies in the HPC co-design space
Exploring emerging technologies in the HPC co-design spaceExploring emerging technologies in the HPC co-design space
Exploring emerging technologies in the HPC co-design spacejsvetter
 
MIG 5th Data Centre Summit 2016 PTS Presentation v1
MIG 5th Data Centre Summit 2016 PTS Presentation v1MIG 5th Data Centre Summit 2016 PTS Presentation v1
MIG 5th Data Centre Summit 2016 PTS Presentation v1blewington
 
PLNOG 8: Ivan Pepelnjak - Data Center Fabrics - What Really Matters
PLNOG 8: Ivan Pepelnjak - Data Center Fabrics - What Really Matters PLNOG 8: Ivan Pepelnjak - Data Center Fabrics - What Really Matters
PLNOG 8: Ivan Pepelnjak - Data Center Fabrics - What Really Matters PROIDEA
 
Data Networks: Next-Generation Optical Access toward 10 Gb/s Everywhere
Data Networks: Next-Generation Optical Access toward 10 Gb/s EverywhereData Networks: Next-Generation Optical Access toward 10 Gb/s Everywhere
Data Networks: Next-Generation Optical Access toward 10 Gb/s EverywhereXi'an Jiaotong-Liverpool University
 
The OptIPuter as a Prototype for CalREN-XD
The OptIPuter as a Prototype for CalREN-XDThe OptIPuter as a Prototype for CalREN-XD
The OptIPuter as a Prototype for CalREN-XDLarry Smarr
 

Similar to Simulating the Universe from quarks to galaxy clusters with DiRAC HPC (20)

TeraGrid Communication and Computation
TeraGrid Communication and ComputationTeraGrid Communication and Computation
TeraGrid Communication and Computation
 
Dcn invited ecoc2018_short
Dcn invited ecoc2018_shortDcn invited ecoc2018_short
Dcn invited ecoc2018_short
 
Accelerating TensorFlow with RDMA for high-performance deep learning
Accelerating TensorFlow with RDMA for high-performance deep learningAccelerating TensorFlow with RDMA for high-performance deep learning
Accelerating TensorFlow with RDMA for high-performance deep learning
 
Manycores for the Masses
Manycores for the MassesManycores for the Masses
Manycores for the Masses
 
Panel: NRP Science Impacts​
Panel: NRP Science Impacts​Panel: NRP Science Impacts​
Panel: NRP Science Impacts​
 
The Coming Age of Extreme Heterogeneity in HPC
The Coming Age of Extreme Heterogeneity in HPCThe Coming Age of Extreme Heterogeneity in HPC
The Coming Age of Extreme Heterogeneity in HPC
 
The SKA Project - The World's Largest Streaming Data Processor
The SKA Project - The World's Largest Streaming Data ProcessorThe SKA Project - The World's Largest Streaming Data Processor
The SKA Project - The World's Largest Streaming Data Processor
 
Barcelona Supercomputing Center, Generador de Riqueza
Barcelona Supercomputing Center, Generador de RiquezaBarcelona Supercomputing Center, Generador de Riqueza
Barcelona Supercomputing Center, Generador de Riqueza
 
Linac Coherent Light Source (LCLS) Data Transfer Requirements
Linac Coherent Light Source (LCLS) Data Transfer RequirementsLinac Coherent Light Source (LCLS) Data Transfer Requirements
Linac Coherent Light Source (LCLS) Data Transfer Requirements
 
Design installation-commissioning-red raider-cluster-ttu
Design installation-commissioning-red raider-cluster-ttuDesign installation-commissioning-red raider-cluster-ttu
Design installation-commissioning-red raider-cluster-ttu
 
DARPA ERI Summit 2018: The End of Moore’s Law & Faster General Purpose Comput...
DARPA ERI Summit 2018: The End of Moore’s Law & Faster General Purpose Comput...DARPA ERI Summit 2018: The End of Moore’s Law & Faster General Purpose Comput...
DARPA ERI Summit 2018: The End of Moore’s Law & Faster General Purpose Comput...
 
Greendroid ppt
Greendroid pptGreendroid ppt
Greendroid ppt
 
Java Thread and Process Performance for Parallel Machine Learning on Multicor...
Java Thread and Process Performance for Parallel Machine Learning on Multicor...Java Thread and Process Performance for Parallel Machine Learning on Multicor...
Java Thread and Process Performance for Parallel Machine Learning on Multicor...
 
Exploring emerging technologies in the HPC co-design space
Exploring emerging technologies in the HPC co-design spaceExploring emerging technologies in the HPC co-design space
Exploring emerging technologies in the HPC co-design space
 
MIG 5th Data Centre Summit 2016 PTS Presentation v1
MIG 5th Data Centre Summit 2016 PTS Presentation v1MIG 5th Data Centre Summit 2016 PTS Presentation v1
MIG 5th Data Centre Summit 2016 PTS Presentation v1
 
PLNOG 8: Ivan Pepelnjak - Data Center Fabrics - What Really Matters
PLNOG 8: Ivan Pepelnjak - Data Center Fabrics - What Really Matters PLNOG 8: Ivan Pepelnjak - Data Center Fabrics - What Really Matters
PLNOG 8: Ivan Pepelnjak - Data Center Fabrics - What Really Matters
 
Data Networks: Next-Generation Optical Access toward 10 Gb/s Everywhere
Data Networks: Next-Generation Optical Access toward 10 Gb/s EverywhereData Networks: Next-Generation Optical Access toward 10 Gb/s Everywhere
Data Networks: Next-Generation Optical Access toward 10 Gb/s Everywhere
 
Umit hw6
Umit hw6Umit hw6
Umit hw6
 
The OptIPuter as a Prototype for CalREN-XD
The OptIPuter as a Prototype for CalREN-XDThe OptIPuter as a Prototype for CalREN-XD
The OptIPuter as a Prototype for CalREN-XD
 
Grid computing & its applications
Grid computing & its applicationsGrid computing & its applications
Grid computing & its applications
 

More from inside-BigData.com

Preparing to program Aurora at Exascale - Early experiences and future direct...
Preparing to program Aurora at Exascale - Early experiences and future direct...Preparing to program Aurora at Exascale - Early experiences and future direct...
Preparing to program Aurora at Exascale - Early experiences and future direct...inside-BigData.com
 
Transforming Private 5G Networks
Transforming Private 5G NetworksTransforming Private 5G Networks
Transforming Private 5G Networksinside-BigData.com
 
The Incorporation of Machine Learning into Scientific Simulations at Lawrence...
The Incorporation of Machine Learning into Scientific Simulations at Lawrence...The Incorporation of Machine Learning into Scientific Simulations at Lawrence...
The Incorporation of Machine Learning into Scientific Simulations at Lawrence...inside-BigData.com
 
How to Achieve High-Performance, Scalable and Distributed DNN Training on Mod...
How to Achieve High-Performance, Scalable and Distributed DNN Training on Mod...How to Achieve High-Performance, Scalable and Distributed DNN Training on Mod...
How to Achieve High-Performance, Scalable and Distributed DNN Training on Mod...inside-BigData.com
 
Evolving Cyberinfrastructure, Democratizing Data, and Scaling AI to Catalyze ...
Evolving Cyberinfrastructure, Democratizing Data, and Scaling AI to Catalyze ...Evolving Cyberinfrastructure, Democratizing Data, and Scaling AI to Catalyze ...
Evolving Cyberinfrastructure, Democratizing Data, and Scaling AI to Catalyze ...inside-BigData.com
 
HPC Impact: EDA Telemetry Neural Networks
HPC Impact: EDA Telemetry Neural NetworksHPC Impact: EDA Telemetry Neural Networks
HPC Impact: EDA Telemetry Neural Networksinside-BigData.com
 
Biohybrid Robotic Jellyfish for Future Applications in Ocean Monitoring
Biohybrid Robotic Jellyfish for Future Applications in Ocean MonitoringBiohybrid Robotic Jellyfish for Future Applications in Ocean Monitoring
Biohybrid Robotic Jellyfish for Future Applications in Ocean Monitoringinside-BigData.com
 
Machine Learning for Weather Forecasts
Machine Learning for Weather ForecastsMachine Learning for Weather Forecasts
Machine Learning for Weather Forecastsinside-BigData.com
 
HPC AI Advisory Council Update
HPC AI Advisory Council UpdateHPC AI Advisory Council Update
HPC AI Advisory Council Updateinside-BigData.com
 
Fugaku Supercomputer joins fight against COVID-19
Fugaku Supercomputer joins fight against COVID-19Fugaku Supercomputer joins fight against COVID-19
Fugaku Supercomputer joins fight against COVID-19inside-BigData.com
 
Energy Efficient Computing using Dynamic Tuning
Energy Efficient Computing using Dynamic TuningEnergy Efficient Computing using Dynamic Tuning
Energy Efficient Computing using Dynamic Tuninginside-BigData.com
 
HPC at Scale Enabled by DDN A3i and NVIDIA SuperPOD
HPC at Scale Enabled by DDN A3i and NVIDIA SuperPODHPC at Scale Enabled by DDN A3i and NVIDIA SuperPOD
HPC at Scale Enabled by DDN A3i and NVIDIA SuperPODinside-BigData.com
 
Versal Premium ACAP for Network and Cloud Acceleration
Versal Premium ACAP for Network and Cloud AccelerationVersal Premium ACAP for Network and Cloud Acceleration
Versal Premium ACAP for Network and Cloud Accelerationinside-BigData.com
 
Zettar: Moving Massive Amounts of Data across Any Distance Efficiently
Zettar: Moving Massive Amounts of Data across Any Distance EfficientlyZettar: Moving Massive Amounts of Data across Any Distance Efficiently
Zettar: Moving Massive Amounts of Data across Any Distance Efficientlyinside-BigData.com
 
Scaling TCO in a Post Moore's Era
Scaling TCO in a Post Moore's EraScaling TCO in a Post Moore's Era
Scaling TCO in a Post Moore's Erainside-BigData.com
 
CUDA-Python and RAPIDS for blazing fast scientific computing
CUDA-Python and RAPIDS for blazing fast scientific computingCUDA-Python and RAPIDS for blazing fast scientific computing
CUDA-Python and RAPIDS for blazing fast scientific computinginside-BigData.com
 
Introducing HPC with a Raspberry Pi Cluster
Introducing HPC with a Raspberry Pi ClusterIntroducing HPC with a Raspberry Pi Cluster
Introducing HPC with a Raspberry Pi Clusterinside-BigData.com
 

More from inside-BigData.com (20)

Major Market Shifts in IT
Major Market Shifts in ITMajor Market Shifts in IT
Major Market Shifts in IT
 
Preparing to program Aurora at Exascale - Early experiences and future direct...
Preparing to program Aurora at Exascale - Early experiences and future direct...Preparing to program Aurora at Exascale - Early experiences and future direct...
Preparing to program Aurora at Exascale - Early experiences and future direct...
 
Transforming Private 5G Networks
Transforming Private 5G NetworksTransforming Private 5G Networks
Transforming Private 5G Networks
 
The Incorporation of Machine Learning into Scientific Simulations at Lawrence...
The Incorporation of Machine Learning into Scientific Simulations at Lawrence...The Incorporation of Machine Learning into Scientific Simulations at Lawrence...
The Incorporation of Machine Learning into Scientific Simulations at Lawrence...
 
How to Achieve High-Performance, Scalable and Distributed DNN Training on Mod...
How to Achieve High-Performance, Scalable and Distributed DNN Training on Mod...How to Achieve High-Performance, Scalable and Distributed DNN Training on Mod...
How to Achieve High-Performance, Scalable and Distributed DNN Training on Mod...
 
Evolving Cyberinfrastructure, Democratizing Data, and Scaling AI to Catalyze ...
Evolving Cyberinfrastructure, Democratizing Data, and Scaling AI to Catalyze ...Evolving Cyberinfrastructure, Democratizing Data, and Scaling AI to Catalyze ...
Evolving Cyberinfrastructure, Democratizing Data, and Scaling AI to Catalyze ...
 
HPC Impact: EDA Telemetry Neural Networks
HPC Impact: EDA Telemetry Neural NetworksHPC Impact: EDA Telemetry Neural Networks
HPC Impact: EDA Telemetry Neural Networks
 
Biohybrid Robotic Jellyfish for Future Applications in Ocean Monitoring
Biohybrid Robotic Jellyfish for Future Applications in Ocean MonitoringBiohybrid Robotic Jellyfish for Future Applications in Ocean Monitoring
Biohybrid Robotic Jellyfish for Future Applications in Ocean Monitoring
 
Machine Learning for Weather Forecasts
Machine Learning for Weather ForecastsMachine Learning for Weather Forecasts
Machine Learning for Weather Forecasts
 
HPC AI Advisory Council Update
HPC AI Advisory Council UpdateHPC AI Advisory Council Update
HPC AI Advisory Council Update
 
Fugaku Supercomputer joins fight against COVID-19
Fugaku Supercomputer joins fight against COVID-19Fugaku Supercomputer joins fight against COVID-19
Fugaku Supercomputer joins fight against COVID-19
 
Energy Efficient Computing using Dynamic Tuning
Energy Efficient Computing using Dynamic TuningEnergy Efficient Computing using Dynamic Tuning
Energy Efficient Computing using Dynamic Tuning
 
HPC at Scale Enabled by DDN A3i and NVIDIA SuperPOD
HPC at Scale Enabled by DDN A3i and NVIDIA SuperPODHPC at Scale Enabled by DDN A3i and NVIDIA SuperPOD
HPC at Scale Enabled by DDN A3i and NVIDIA SuperPOD
 
State of ARM-based HPC
State of ARM-based HPCState of ARM-based HPC
State of ARM-based HPC
 
Versal Premium ACAP for Network and Cloud Acceleration
Versal Premium ACAP for Network and Cloud AccelerationVersal Premium ACAP for Network and Cloud Acceleration
Versal Premium ACAP for Network and Cloud Acceleration
 
Zettar: Moving Massive Amounts of Data across Any Distance Efficiently
Zettar: Moving Massive Amounts of Data across Any Distance EfficientlyZettar: Moving Massive Amounts of Data across Any Distance Efficiently
Zettar: Moving Massive Amounts of Data across Any Distance Efficiently
 
Scaling TCO in a Post Moore's Era
Scaling TCO in a Post Moore's EraScaling TCO in a Post Moore's Era
Scaling TCO in a Post Moore's Era
 
CUDA-Python and RAPIDS for blazing fast scientific computing
CUDA-Python and RAPIDS for blazing fast scientific computingCUDA-Python and RAPIDS for blazing fast scientific computing
CUDA-Python and RAPIDS for blazing fast scientific computing
 
Introducing HPC with a Raspberry Pi Cluster
Introducing HPC with a Raspberry Pi ClusterIntroducing HPC with a Raspberry Pi Cluster
Introducing HPC with a Raspberry Pi Cluster
 
Overview of HPC Interconnects
Overview of HPC InterconnectsOverview of HPC Interconnects
Overview of HPC Interconnects
 

Recently uploaded

How to Remove Document Management Hurdles with X-Docs?
How to Remove Document Management Hurdles with X-Docs?How to Remove Document Management Hurdles with X-Docs?
How to Remove Document Management Hurdles with X-Docs?XfilesPro
 
Beyond Boundaries: Leveraging No-Code Solutions for Industry Innovation
Beyond Boundaries: Leveraging No-Code Solutions for Industry InnovationBeyond Boundaries: Leveraging No-Code Solutions for Industry Innovation
Beyond Boundaries: Leveraging No-Code Solutions for Industry InnovationSafe Software
 
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
08448380779 Call Girls In Diplomatic Enclave Women Seeking MenDelhi Call girls
 
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024BookNet Canada
 
Factors to Consider When Choosing Accounts Payable Services Providers.pptx
Factors to Consider When Choosing Accounts Payable Services Providers.pptxFactors to Consider When Choosing Accounts Payable Services Providers.pptx
Factors to Consider When Choosing Accounts Payable Services Providers.pptxKatpro Technologies
 
SQL Database Design For Developers at php[tek] 2024
SQL Database Design For Developers at php[tek] 2024SQL Database Design For Developers at php[tek] 2024
SQL Database Design For Developers at php[tek] 2024Scott Keck-Warren
 
Azure Monitor & Application Insight to monitor Infrastructure & Application
Azure Monitor & Application Insight to monitor Infrastructure & ApplicationAzure Monitor & Application Insight to monitor Infrastructure & Application
Azure Monitor & Application Insight to monitor Infrastructure & ApplicationAndikSusilo4
 
Unblocking The Main Thread Solving ANRs and Frozen Frames
Unblocking The Main Thread Solving ANRs and Frozen FramesUnblocking The Main Thread Solving ANRs and Frozen Frames
Unblocking The Main Thread Solving ANRs and Frozen FramesSinan KOZAK
 
Pigging Solutions in Pet Food Manufacturing
Pigging Solutions in Pet Food ManufacturingPigging Solutions in Pet Food Manufacturing
Pigging Solutions in Pet Food ManufacturingPigging Solutions
 
Handwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed textsHandwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed textsMaria Levchenko
 
Presentation on how to chat with PDF using ChatGPT code interpreter
Presentation on how to chat with PDF using ChatGPT code interpreterPresentation on how to chat with PDF using ChatGPT code interpreter
Presentation on how to chat with PDF using ChatGPT code interpreternaman860154
 
Scaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organizationScaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organizationRadu Cotescu
 
IAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI SolutionsIAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI SolutionsEnterprise Knowledge
 
A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)Gabriella Davis
 
From Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time AutomationFrom Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time AutomationSafe Software
 
04-2024-HHUG-Sales-and-Marketing-Alignment.pptx
04-2024-HHUG-Sales-and-Marketing-Alignment.pptx04-2024-HHUG-Sales-and-Marketing-Alignment.pptx
04-2024-HHUG-Sales-and-Marketing-Alignment.pptxHampshireHUG
 
My Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 PresentationMy Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 PresentationRidwan Fadjar
 
08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking Men08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking MenDelhi Call girls
 
[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdf[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdfhans926745
 
Maximizing Board Effectiveness 2024 Webinar.pptx
Maximizing Board Effectiveness 2024 Webinar.pptxMaximizing Board Effectiveness 2024 Webinar.pptx
Maximizing Board Effectiveness 2024 Webinar.pptxOnBoard
 

Recently uploaded (20)

How to Remove Document Management Hurdles with X-Docs?
How to Remove Document Management Hurdles with X-Docs?How to Remove Document Management Hurdles with X-Docs?
How to Remove Document Management Hurdles with X-Docs?
 
Beyond Boundaries: Leveraging No-Code Solutions for Industry Innovation
Beyond Boundaries: Leveraging No-Code Solutions for Industry InnovationBeyond Boundaries: Leveraging No-Code Solutions for Industry Innovation
Beyond Boundaries: Leveraging No-Code Solutions for Industry Innovation
 
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
 
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
 
Factors to Consider When Choosing Accounts Payable Services Providers.pptx
Factors to Consider When Choosing Accounts Payable Services Providers.pptxFactors to Consider When Choosing Accounts Payable Services Providers.pptx
Factors to Consider When Choosing Accounts Payable Services Providers.pptx
 
SQL Database Design For Developers at php[tek] 2024
SQL Database Design For Developers at php[tek] 2024SQL Database Design For Developers at php[tek] 2024
SQL Database Design For Developers at php[tek] 2024
 
Azure Monitor & Application Insight to monitor Infrastructure & Application
Azure Monitor & Application Insight to monitor Infrastructure & ApplicationAzure Monitor & Application Insight to monitor Infrastructure & Application
Azure Monitor & Application Insight to monitor Infrastructure & Application
 
Unblocking The Main Thread Solving ANRs and Frozen Frames
Unblocking The Main Thread Solving ANRs and Frozen FramesUnblocking The Main Thread Solving ANRs and Frozen Frames
Unblocking The Main Thread Solving ANRs and Frozen Frames
 
Pigging Solutions in Pet Food Manufacturing
Pigging Solutions in Pet Food ManufacturingPigging Solutions in Pet Food Manufacturing
Pigging Solutions in Pet Food Manufacturing
 
Handwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed textsHandwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed texts
 
Presentation on how to chat with PDF using ChatGPT code interpreter
Presentation on how to chat with PDF using ChatGPT code interpreterPresentation on how to chat with PDF using ChatGPT code interpreter
Presentation on how to chat with PDF using ChatGPT code interpreter
 
Scaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organizationScaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organization
 
IAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI SolutionsIAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI Solutions
 
A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)
 
From Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time AutomationFrom Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time Automation
 
04-2024-HHUG-Sales-and-Marketing-Alignment.pptx
04-2024-HHUG-Sales-and-Marketing-Alignment.pptx04-2024-HHUG-Sales-and-Marketing-Alignment.pptx
04-2024-HHUG-Sales-and-Marketing-Alignment.pptx
 
My Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 PresentationMy Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 Presentation
 
08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking Men08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking Men
 
[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdf[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdf
 
Maximizing Board Effectiveness 2024 Webinar.pptx
Maximizing Board Effectiveness 2024 Webinar.pptxMaximizing Board Effectiveness 2024 Webinar.pptx
Maximizing Board Effectiveness 2024 Webinar.pptx
 

Simulating the Universe from quarks to galaxy clusters with DiRAC HPC

  • 1. 40 powers of 10 Simulating the Universe from quarks to galaxy clusters with the DiRAC HPC Facility Dr Mark Wilkinson University of Leicester Director, DiRAC HPC Facility DiRAC
  • 2. DiRAC Distributed HPC Facility for theoretical astrophysics, particle physics, cosmology and nuclear physics Combined: • ~90,000 cores • ~5 Pflop/s • >10 PB storage Extreme Scaling (Edinburgh) Data Intensive (Cambridge) Data Intensive (Leicester) Memory Intensive (Durham)
  • 3. The DiRAC Facility, a brief history 2009: DiRAC-1: systems installed at 13 host sites Nov 2011: DiRAC-2 awarded £15M capital by BIS - 5 systems at 4 sites - Cambridge, Durham, Edinburgh, Leicester - Procurement completed in 100 days Dec 2012: DiRAC-2 operations begin - Systems full within 1 week - usage >90% - Access via international peer-review process - Free to use for STFC-funded researchers April 2017: DiRAC-2.5 operations begin - 3 services: Extreme Scaling, Memory Intensive, Data Intensive April 2018: DiRAC-2.5x - interim funding to replace 2012 hardware and support 2018/19 science programme Dec 2018: DiRAC-2.5y - BEIS-funded stop-gap 2x upgrade of all DiRAC services
  • 4. • Direct numerical simulation and modelling is core theory research activity • HPC systems are main scientific instruments for theory research • Computational requirements of models increasing due to • Increased resolution: running models with existing physics at finer scales. • Increased complexity: introducing new physics into models to reflect progress in theoretical understanding; often needed to match resolution • Coupling of models: multi-physics, multi-scale modelling; • Quantification of modelling uncertainty using large ensembles of simulations to provide robust statistics • Constant process of refining and re-defining our tools • Growing requirement for simulations and modelling concurrent with observations so that models evolve in line with data acquired • Observational facilities need access to significant local computing capabilities as well as option to burst out to the larger, national facilities. Computing in theory research
  • 5. The DiRAC approach to service design Science case Workflow assessment Technical case Peer-reviewed, scientific justification for resources Peer-reviewed, high-level technical specifications Technical design Co-design with industry partners Individual service specifications
  • 6. Diverse science cases require heterogenous architectures Extreme Scaling “Tesseract” (Edinburgh) Memory Intensive “COSMA” (Durham) Data Intensive “DIaL” and “CSD3” (Leicester & Cambridge) 230 TB RAM to support largest cosmological simulations Heterogeneous architecture to support complex simulation and modelling workflows 2 Pflop/s to support largest lattice-QCD simulations DiRAC
  • 7. scientific impact is highly significant, sustained and of world leading quality. Figure 3-7: Facility Citations and Paper Record 2015-2018 3.5 WP5: Project Management DiRAC Science Impact Analysis • Refereed publications since 2012: > 950 papers; > 49000 citations • Second-most cited paper in all astronomy in 2015 • 199 refereed papers in 2017
  • 8. DiRAC@Edinburgh – Extreme Scaling (ES) • HPE SGI 8600 • >1.8 PF • 1468 standard compute nodes • Two 2.1GHz 12-core Intel Xeon Skylake 4116 processors • 96GB RAM • 8 GPU compute nodes • Two 2.1GHz 12-core Intel Xeon Skylake 4116 processors • 96 GB RAM • 4 NVidia V100 (Volta) GPU accelerators connected over PCIe • Hypercube Intel Omnipath interconnect • 3PB Lustre DDN storage “Tesseract”
  • 9. The forces of nature
  • 10. Inside the proton Image: Arpad Horvath - Own work, CC BY-SA 2.5, https://commons.wikimedia.org/w/index.php?curid=637353 quark Strong interactions carried by gluons
  • 11. Understanding the strong force is key to testing the Standard Modal of Particle Physics - binds quarks into hadrons that we see in experiment ATLAS@Large Hadron Collider (LHC) Connecting observed hadron properties to those of quarks requires full non-perturbative treatment of Quantum Chromodynamics - lattice QCD DALI Run=16449 Evt=4055 PH 1cm| IP B D ππ KK K π e− − − + + +s s Calculations of hadron masses and rates for simple weak decays allow tests of Standard Model
  • 12. • L4 local volume (space + time) • finite di↵erence operator 8 point stencil • ⇠ 1 L of data references come from o↵ node Scaling QCD sparse matrix requires interconnect bandwidth for halo exchange Bnetwork ⇠ Bmemory L ⇥R where R is the reuse factor obtained for the stencil in caches • Aim: Distribute 1004 datapoints over 104 nodes Scalable QCD: sparse matrix PDE solver communications (Boyle et al.)
  • 13. Immediate Big HPC roadmap 0.1 1 10 100 1000 2010 2012 2014 2016 2018 2020 2022 PF/system IBM-BGQ/Mira Intel-Broadwell/Cori1 Intel-KNL/Cori2 Intel-KNL/Oakforest PACS IBM/Nvidia/Summit/Sierra Intel-KNH/Aurora Fujitsu-ARM/post-K Frontier/ORNL 0.1 1 10 100 1000 2010 2012 2014 2016 2018 2020 2022 SPTF/node IBM-BGQ/Mira Intel-Broadwell/Cori1 Intel-KNL/Cori2 Intel-KNL/Oakforest PACS IBM/Nvidia/Summit 0.1 1 10 100 1000 2010 2012 2014 2016 2018 2020 2022 commsGB/s/node IBM-BGQ/Mira Intel-Broadwell/Cori1 Intel-KNL/Cori2 Intel-KNL/Oakforest PACS IBM/Nvidia/Summit/Sierra • 400x+ increase in SP node performance accompanied by NO increase in interconnect Immediate “Big HPC” roadmap
  • 14. Grid QCD code Design considerations • Performance portable across multi and many core CPU’s SIMD⌦OpenMP⌦MPI • Performance portable to GPU’s SIMT⌦offload⌦MPI • N-dimensional cartesian arrays • Multiple grids • Data parallel C++ layer : Connection Machine inspired Grid QCD code (Boyle et al.)
  • 15. • Relatively cheap node: high node count and scalability • Improve price per core, bandwidth per core, reduce power Tesseract performance per node vs nodes, volume GF/spernode 0.0 150.0 300.0 450.0 600.0 Nodes 1 16 32 64 128 256 12^4 16^4 24^4 nodes (single switch) delivers bidirectional 25GB/s to every node (wirespeed) 2 nodes topology aware bidirectional 19GB/s • 76% wirespeed using every link in system concurrently DiRAC HPE ICE-XA hypercube network • Edinburgh HPE 8600 system (Installed March 2018, 2 days ahead of schedule) • Low end Skylake Silver 4116, 12 core parts • Single rail Omnipath interconnect • Relatively cheap node: high node count and scalability • Improve price per core, bandwidth per core, reduce power Tesseract performance per node vs nodes, volume GF/spernode 0.0 150.0 300.0 450.0 600.0 Nodes 1 16 32 64 128 256 12^4 16^4 24^4 • 16 nodes (single switch) delivers bidirectional 25GB/s to every node (wirespeed) • 512 nodes topology aware bidirectional 19GB/s • 76% wirespeed using every link in system concurrently • Low end Skylake Silver 4116, 12 core parts • Single rail Omnipath interconnect • Relatively cheap node: high node count and scalability • Improve price per core, bandwidth per core, reduce power Tesseract performance per node vs nodes, volume GF/spernode 0.0 150.0 300.0 450.0 600.0 Nodes 1 16 32 64 128 256 12^4 16^4 24^4 • 16 nodes (single switch) delivers bidirectional 25GB/s to every node (wirespeed) • 512 nodes topology aware bidirectional 19GB/s • 76% wirespeed using every link in system concurrently • Low end Skylake Silver 4116, 12 core parts • Single rail Omnipath interconnect • Relatively cheap node: high node count and scalability • Improve price per core, bandwidth per core, reduce power Tesseract performance per node vs nodes, volume GF/spernode 0.0 150.0 300.0 450.0 600.0 Nodes 1 16 32 64 128 256 12^4 16^4 24^4 16 nodes (single switch) delivers bidirectional 25GB/s to every node (wirespeed) 512 nodes topology aware bidirectional 19GB/s • 76% wirespeed using every link in system concurrently • Small project with SGI/HPE on Mellanox EDR networks (James Southern) • Embed 2n QCD torus inside hypercube so that nearest neigbour comms travels single hop 4x speed up over default MPI Cartesian communicators on large systems ) Customise HPE 8600 (SGI ICE-XA) to use 16 = 24 nodes per leaf switch Boyle et al.
  • 16. Other Turing work • Intel’s US Patent Application 20190042544 (FP16-S7E8 MIXED PRECISION FOR DEEP LEARNING AND OTHER ALGORITHMS) • Authors: Boyle (ATI, Edinburgh), Sid Kashyap, Angus Lepper (Intel, former DiRAC RSE’s) • Systematic study using Gnu Multi Precision library. • BFP16 displays greater numerical stability for machine learning training. • Understanding: histogramme of multiply results during SGD gradient calculations • Patent application full text searchable on uspto.gov DiRAC / Alan Turing Institute / Intel Collaboration
  • 17. Software Innovation - AI on HPC cessors. ������ ����� ���� ��� � �� ��� ���� ����� ������ ����� ����� ����� ������� ����� ����� ���� ����� ����� ���� ������� G. 5: Wall clock time per reduction call vs. vector length before and after our optimisation. The large vector reducti formance is ten times better after our optimisations on large vector lengths. The gain includes both computation accelerati d communication acceleration. the interface to memory management is not ideal. ther, it is likely that method dealloc and alloc routine provided by collectives.h should have been used consistent th use of any other “free” operation declared illegal. Further, this would have enabled simpler modification of t ocation and deallocation implementation. It would also have been better to separate the vector allocation from t duction operation; so that in hot loops the programmer could choose to reuse the same allocation. ur first optimisations were high level: i) remove the expectation that the caller deallocates the returned vector; Boyle et al. 2017 • Demonstration of factor 10 speed-up in the Baidu Research optimised reduction code - a publicly available code designed to optimise the performance limiting steps in distributed machine learning (ML) • potentially disruptive implications for design of cloud systems - shows that ML workflows can achieve 10x performance improvement when fully optimised and implemented on traditional HPC architectures
  • 18. gure 1: Left panel: A ‘fan’ plot for the light quark ud and us decay constants (normalised with the erage ud, us decay constant) for one lattice spacing against the pion mass; Right panel: The heavy b uark ub, sb decay constants, again normalised, for several lattice spacings also plotted against the on mass. ow further extended this programme to the heavier quark masses and in particular the b, [2]. the right panel of the figure we give the equivalent plot for the B and Bs mesons, together ith the present (FLAG16) values. Bornyakov et al. 2017 Hollitt et al. 2018 Testing the standard model of particle physics • Mesons are composed of quark-antiquark pairs • Their decay rates into other particles provide a strong test of the standard model • Anomalies may point to new physics • DiRAC researchers have developed new, computationally efficient technique for estimating decay rates
  • 19. DiRAC@Durham - Memory Intensive (MI) • COSMA6 (2016) • IBM/Lenovo/DDN • 8192 Intel SandyBridge, 2.6 GHz cores • Combined 65 Tbyte of RAM; • Mellanox FDR10 interconnect in 2:1 blocking; • 2.5 Pbyte of Lustre data storage • COSMA7 (2018) • Dell • 23,600 Intel Skylake 4116 cores • Interconnect: Mellanox EDR in a 2:1 blocking configuration with islands of 24 nodes; • Combined 230 Tbyte of RAM; • a fast checkpointing i/o system (343 Tbyte) with peak performance of 185 Gbyte/second write and read; • >4 PB storage Durham – MI m was delivered by Dell in March 2018; y Alces service started on 1 May 2018 engagement: closely with the Industrial Strategy of partment for Business, Energy and ial Strategy (BEIS). g leads to industrial engagement; this in innovation, both leading to ts academia and the wider industry.
  • 20. The Universe, a brief history Credit: NASA / WMAP Science Team DiRAC supports calculations on all physical scales
  • 21. The components of the Universe Credit: NASA / WMAP Science Team & Planck Satellite Team 26.8% 4.9% 68.3%
  • 22. ICC Durham; Virgo Consortium The EAGLE Simulation DiRAC Dark matter (N-body)
  • 23. The EAGLE Simulation ICC Durham; Virgo ConsortiumDiRAC Gas and stars (hydrodynamics)
  • 24. Research Software Engineers “With great power comes great responsibility” • Science requirements for DiRAC-3 demand 10-40x increases in computing power to stay competitive - hardware alone cannot deliver this • We can no longer rely on “free lunch” from the Xeon era • Vectorisation and code efficiency now critical • Next generation hardware is more difficult to program efficiently • RSEs are increasingly important - RSEs can help with code profiling, optimisation, porting, etc • DiRAC now has 3 RSEs: effort allocated via peer review process algorithmic computational David Keyes
  • 25. FirstGalaxies SolarSystemForms Reionization Milky Way 13121110987654321 Age of the Universe (Gyr) Present DayBig Bang The Evolution of the Universe First simulation on DiRAC-2.5y Memory Intensive system at Durham carried out with SWIFT, a revolutionary new cosmological hydrodynamics code developed at Durham which is 20x faster than the state-of-the-art. Key Problems Origin of galaxies Indentity of the dark matter Nature of the dark energy
  • 26. • A fast checkpointing i/o system (343 Tbyte) with peak performance of 185 Gbyte/second write and read • 15 Lustre Object Storage Servers on Dell 640 nodes: • 2 x Intel Skylake 5120 processors • 192 Gbyte of RAM • 8 x 3.2 TB NVMe SFF drives. • 1 Mellanox EDR card • A user code benchmark produced 180 Gbyte/second write and read – this is almost wire-speed! • At time of installation this was the fastest filesystem in production in Europe Memory Intensive: burst buffer for checkpointing
  • 27. 0 500,000 1,000,000 1,500,000 2,000,000 2,500,000 3,000,000 24-hour 12-hour 6-hour 4-hour 2-hour 1-hour KiloWatthoursover5years Axis Title Power usage for snapshots with different performance solutions 30GB/sec 120GB/sec 140GB/sec Memory Intensive: burst buffer for checkpointing Power saving Heck (2018)
  • 28. Core hours lost in checkpointing over 5 years Snapshot period 24-hr 12-hr 6-hr 4-hr 2-hr 1-hr Hours/ snapshot 30GB/sec 7.1M 14.2M 28.4M 42.5M 85.1M 170.1M 0.95 120GB/sec 1.8M 3.5M 7.1M 10.6M 21.3M 42.5M 0.24 140GB/sec 1.5M 3.0M 6.1M 9.1M 18.2M 36.54M 0.20 • Total number of available cpu hours per year: 36M (4116 cores)  • ~13% gain in available core hours due to faster checkpointing Memory Intensive: burst buffer for checkpointing Heck (2018)
  • 29. Uranus collision? 8 J. A. Kegerreis et al. 10 5 0 x Position (R ) 10 5 0 yPosition(R) 10 5 0 x Position (R ) 106 107 108 SpecificInternalEnergyJkg1 Figure 6. A mid-collision snapshots of a grazing impact with 108 SPH particles – compared with the more head-on collision in Fig. 5 – coloured by their material and internal energy, showing some of the detailed evolution and mixing that can now be resolved. In the left panel, light and dark grey show the target’s ice and rock material, respectively, and purple and brown show the same for the impactor. Light blue is the target’s H-He atmosphere. mon aim in cosmology, to simulate a larger patch of the uni- verse), or we can study a small system with higher resolution to model smaller details. not) question that we ask a computer to solve, so this is an important first step. Kegerreis et al. (2019), subm. Planet collisions Giant impact explains Uranus spin axis and cold temperature • 108-particle simulations carried out on DiRAC Memory Intensive service (Durham) using new hydro+Nbody code, Swift • Hardware and software for this calculation developed with DiRAC support.
  • 30. DiRAC & the discovery of Gravitational waves On September 14, 2015 at 09:50:45 UTC, the LIGO Hanford, WA, and Livingston, LA, observatories detected pact binary waveforms [44] recovered GW most significant event from each detector f tions reported here. Occurring within the • Simulations of binary black hole mergers performed on DiRAC DataCentric system (COSMA5) • Crucial for interpretation of LIGO gravitational wave detection Abbott et al., (2016) DiRAC
  • 31. DiRAC@Leicester – Data Intensive (DIaL) • HPE system • 400 compute nodes • 2x Intel Xeon Skylake 6140, 2.3GHz, 18-core processors • Dual FMA AVX512 • 192 GB RAM • 1 x 6TB SuperDome Flex with 144 X6154 cores (3.0GHz) • 3 x 1.5TB fat nodes with 36 X6140 cores • Mellanox EDR interconnect in a 2:1 blocking setup • 3PB Lustre storage. • 150 TB flash storage for data intensive workflows HPE/Arm/Suse Catalyst UK Arm system • 4000-core Thunder X2-based cluster installed in January 2019 • Infiniband all-to-all interconnect
  • 32. Shared platform with local Cambridge and EPSRC Tier 2 clusters • Dell multi-architecture system (Skylake, KNL, Nvidia GPU) • 484 Skylake nodes • 2 x Intel Xeon Skylake 6142 2.6GHz 16-core processors • Mix of 192 GB/node and 384 GB/node • 44 Intel KNL nodes • Intel Xeon Phi CPU 7210 @ 1.30GHz • 96GB of RAM per node • 12 NVIDIA GPU nodes • four NVIDIA Tesla P100 GPUs • 96GB memory per node • connected by Mellanox EDR Infiniband • Intel OmniPath interconnect in 2:1 blocking configuration • 1.5PB of Lustre disk storage; • 0.5PB flash storage-based “Data Accelerator” DiRAC@Cambridge – Data Intensive (DIaC)
  • 34. DiRAC HPC Training • DiRAC provides access to training from wide pool of providers • Currently offering: - DiRAC Essentials Test: now available online (and compulsory!) - Workshops and Hackathons • Coming soon: - Domain-specific workshops - Online individual training portal Why do we do this? - maximise DiRAC science output - flexibility to adopt most cost-effective technologies - future-proofing our software and skills - contributes to increasing skills of wider economy
  • 35. (Better systems) + (Better software) = Better science This requires: • More engagement in hardware and software co-design • Enhanced training and knowledge transfer DiRAC