Lecture01 Intro ToHPC
Lecture01 Intro ToHPC
What is HPC ?
Parallel computers
What do they have in common ?
Where and Why HPC ?
200 km 25 km
Complex problem: climate change over
the Mediterranean sea
• Resolution:
– 200km -> ~ 1GB 2km -> ? GB
200 km 25 km
Complex problems solved by
simulations
● Simulation has become the way to research and
develop new scientific and engineering solutions.
● Used nowadays in leading science domains like
aerospace industry, astrophysics, etc.
● Challenges related to the complexity, scalability
and data production of the simulators arise.
● Impact on the relaying IT infrastructure.
Interested in more example ?
• See chapter one section
1.2 of reference 4
• Look around on the
internet..
Research is changing..
● Inference Spiral of System Science
As models become more complex and new data bring in more
information, we require ever increasing computational power
Data are flooding us..
From https://insidebigdata.com/2019/07/22/converged-hpc-
clusters/
Complex problem 2
Protein folding..
Determine the e structure of the protein from its
aminoacid sequence
- Solve it computationally..
A. Run very long MD simulation on each sequence and wait..
B. Predict new structures training an AI algorithm..
Approach A: Folding@home
• A more than 20 year project
• It allows everybody to run an MD simulation
contributing to the problem..
• Impressive distributed computational power..
• From statistics page:
OS GPU CPU Tflops GPU Tflops x86
• AlphaFold is an AI system
developed by DeepMind that
predicts a protein’s 3D
structure from its amino acid
sequence.
• Presented in 2018 and 2020
at CASP outperformed all
other approaches
• Scores obtained are roughly
equivalent to experimentally
determined structured
AlphaFold DB
AlphaFold DB release in July 2022 open access to over 200
million protein structure predictions to accelerate scientific
research.
We successfully scaled the AlphaFold model training to 512 NVIDIA A100 GPUs and
obtained aggregate 6.02 PetaFLOPs at the training stage. The overall training time is
reduced to 67 hours from 11 days with significant economic cost savings.
Agenda
What is HPC ?
Parallel Computers
HPC: a first second definition
High Performance Computing (HPC) is the use of
servers, clusters, and supercomputers – plus
associated software, tools, components, storage,
and services – for scientific, engineering, or
analytical tasks that are particularly intensive in
computation, memory usage, or data management
MIXING/INTERPLAYING ROLES
INCREASES COMPETENCE LEVELS
Yet another definition
• HPC incorporates all
facets of three disciplines:
– Technology
– Methodology
– Application
What is HPC ?
Parallel Computers
It is all about Performance
• It is difficult to define Performance properly
“speed” / “how fast” are vague terms
• Performance as a measure again ambiguous
and not clearly defined and in its
interpretation
• In any case performance it is at core to HPC
as a discipline
• Let discuss it in some details
Does P stand just for Performance ?
• Performance is not always what matters..
to reflect a greater focus on the productivity,
rather than just the performance, of large-scale
computing systems, many believe that HPC
should now stand for High Productivity
Computing. [ from wikipedia]
FLOPS=clock_rate*Number_of_FP_operation*Number_of_cores
Sustained (peak) performance
• Real (sustained) performance: a measure
FLOPS= (total number of floating point operations done by a
program) / (time the program takes to run in second)
What is HPC ?
Parallel Computers
TOP 500 List
What is HPC ?
Parallel Computers
• To be continued