Discover EuroHPC JU

EuroHPC JU

The European High Performance Computing Joint Undertaking (EuroHPC JU) is a legal and funding entity, created in 2018 and located in Luxembourg.

The EuroHPC Joint Undertaking allows the EU and EuroHPC participating countries to coordinate their efforts and pool their resources with the objective of deploying world-class exascale supercomputers in Europe, able to perform more than one trillion (1018operations per second and develop innovative supercomputing technologies and applications . By making Europe a world leader in high performance computing (HPC), the EuroHPC JU seeks to provide computing solutions, improve cooperation in advanced scientific research, boost industrial competitiveness, and ensure European technological and digital autonomy.

Currently, the Joint Undertaking is supporting the following activities:

Developing a world-class supercomputing infrastructure: procuring and deploying by 2021 in the EU three pre-exascale supercomputers (capable of at least 1017 calculations per second) and five petascale supercomputers (capable of at least 1015 calculations per second). These new machines will be located across the European Union and will be available to Europe's private and public users, scientific and industrial users everywhere in Europe. 

Once all the procurement processes are completed, the three pre-exascale supercomputers will be located at the following supercomputing centres:

While the five petascale supercomputers will be located in the following supercomputing centres:

Supporting research and innovation activities: developing and maintaining an innovative European supercomputing ecosystem, stimulating a technology supply industry (from low-power processors to software and middleware, and their integration into supercomputing systems), and making supercomputing resources in many application areas available to a large number of public and private users, including small and medium-sized enterprises. 

Through its research and innovation agenda, the EuroHPC JU is also strengthening the European knowledge base in HPC technologies and bridging the digital skills gap, notably through the creation of a network of national HPC Competence Centres. The Competence Centres will act locally to ease access to European HPC opportunities in different industrial sectors, delivering tailored solutions for a wide variety of users.

 

Members

The EuroHPC Joint Undertaking is composed of public and private members:

Public members:

  • the European Union (represented by the Commission),
  • Member States and Associated Countries that have chosen to become members of the Joint Undertaking: Austria, Belgium, Bulgaria, Croatia, Cyprus, Czech Republic, Denmark, Estonia, Finland, France, Germany, Greece, Hungary, Iceland, Ireland, Italy, Latvia, Lithuania, Luxembourg, Malta,  Montenegro, the Netherlands, North Macedonia, Norway, Poland, Portugal, Romania, Slovakia, Slovenia, Spain, Sweden, Switzerland, and Turkey.

Private members:

EuroHPC JU members map

Supercomputers

Currently seven EuroHPC supercomputers are under construction across Europe:

LUMI

Visual of the LUMI machine, based on a HPE Cray EX Supercomputer

© HPE

The LUMI system will be a Cray EX supercomputer supplied by Hewlett Packard Enterprise (HPE) and located in Finland. 

Sustained performance: 375 petaflops
Peak performance: 552 petaflops
Compute partitions: GPU partition (LUMI-G), x86 CPU-partition (LUMI-C), data analytics partition (LUMI-D), container cloud partition (LUMI-K)
Central Processing Unit (CPU): The LUMI-C partition will feature 64-core next-generation AMD EPYC™ CPUs
Graphics Processing Unit (GPU): LUMI-G based on the future generation AMD Instinct™ GPU
Storage capacity: LUMI’s storage system will consist of three components. First, there will be a 7-petabyte partition of ultra-fast flash storage, combined with a more traditional 80-petabyte capacity storage, based on the Lustre parallel filesystem, as well as a data management service, based on Ceph and being 30 petabytes in volume. In total, LUMI will have a storage of 117 petabytes and a maximum I/O bandwidth of 2 terabytes per second
Applications:  AI,  especially deep learning, and traditional large scale simulations combined with massive scale data analytics in solving one research problem
Other details:

LUMI takes over 150m2 of space, which is about the size of a tennis court. The weight of the system is nearly 150 000 kilograms (150 metric tons)

 

LEONARDO

Visual of LEONARDO based on a BullSequana XH2000 from ATOS

© Atos

Leonardo will be supplied by ATOS, based on BullSequana XH2000 supercomputer and located in Italy. 

Sustained performance:  249.4 petaflops
Peak performance: 322.6 petaflops
Compute partitions Booster, hybrid CPU-GPU module delivering 240 PFlops, Data-Centric, delivering 9 Pflops and featuring DDR5 Memory and local NVM for data analysis
Central Processing Unit (CPU): Intel Ice-Lake (Booster), Intel Sapphire Rapids (data-centric)
Graphics Processing Unit (GPU): NVIDIA Ampere architecture-based GPUs, delivering 10 exaflops of FP16 Tensor Flow AI performance
Storage capacity: Leonardo is equipped with over 100 petabytes of state-of-the-art storage capacity and 5PB of High Performance storage
Applications: The system targets: modular computing, scalable computing applications, data-analysis computing applications, visualization applications and interactive computing applications, urgent and cloud computing
Other details: Leonardo will be hosted in the premises of the Tecnopolo di Bologna. The area devoted to the EuroHPC Leonardo system includes 890 sqm of data hall, 350 sqm of data storage, electrical and cooling and ventilation systems, offices and ancillary spaces

 

MELUXINA

 

MeluXina supercomputer

 

MeluXina will be supplied by Atos, based on the BullSequana XH2000 supercomputer platform and located in Luxembourg. 

Sustained performance: Committed 10 petaflops HPL (Accelerator - GPU Module), 2+ petaflops HPL (Cluster Module)
Peak performance: Expected 15+ petaflops HPL and ~500 petaflops AI (Accelerator - GPU Module), 3+ petaflops HPL (Cluster Module)
Compute partitions: Cluster, Accelerator - GPU, Accelerator - FPGA, Large Memory
Central Processing Unit (CPU): AMD EPYC
Graphics Processing Unit (GPU): NVIDIA Ampere A100
Storage capacity: 20 petabytes main storage with an all-flash scratch tier at 400GB/s, and a 5 petabytes tape library expandable to 100 petabytes
Applications: Traditional Computational, AI and Big Data/HPDA workloads
Other details: Modular Supercomputer Architecture with a Cloud Module for complex use cases and persistent services, an aggregated 476TB RAM, Infiniband HDR interconnect in Dragonfly+ topology, high speed links to the GÉANT network and Public Internet

 

VEGA

 

Vega HPC

© Atos

Vega was supplied by Atos, based on an BullSequana XH2000 supercomputer and located in Slovenia.

Sustained performance: 6,9 petaflops
Peak performance: 10,1 petaflops
Compute partitions: CPU partition: 960 nodes, 256GB memory/node, 20% double memory, HDR100 & GPU partition: 60 nodes, HDR200
Central Processing Unit (CPU) : 122.800 cores, 1920 CPUs, AMD Epyc 7H12
Graphics Processing Unit (GPU): 240 Nvidia A100 cards
Storage capacity: High-performance NVMe Lustre (1PB), large-capacity Ceph (23PB)
Applications: Traditional Computational, AI, Big Data/HPDA, Large-scale data processing
Other details: Wide bandwidth for data transfers to other national and international computing centres (up to 500 Gbit/s). Data processing throughput 400GB/s from high-performance storage and 200Gb/s from large capacity storage

 

KAROLINA

 

HPC Karolina picture

 

Karolina will be supplied by Hewlett Packard Enterprise (HPE), based on an HPE Apollo 2000Gen10 Plus and HPE Apollo 6500 supercomputers and located in the Czech Republic.

Sustained performance: 9,13 petaflops
Peak performance: 15.2 petaflops
Compute partitions:

The supercomputer will consist of 6 main parts:  

  • a universal part for standard numerical simulations, which will consist of approximately 720 computer servers with a theoretical peak performance of 3.8 PFlop/s,
  • an accelerated part with 70 servers and each of them being equipped with 8 GPU accelerators providing a performance of 11 PFlop/s for standard HPC simulations and up to 150 PFlop/s for artificial intelligence computations, 
  • a part designated for large dataset processing that will provide a shared memory of as high as 24 TB, and a performance of 74 TFlop/s,
  • 36 servers with a performance of 131 TFlop/s will be dedicated for providing cloud services,
  • a high-speed network to connect all parts as well as individual servers at a speed of up to 200 Gb/s,
  • data storages that will provide space for more than 1 PB of user data and will also include high-speed data storage with a speed of 1 TB/s for simulations as well as computations in the fields of advanced data analysis and artificial intelligence.
Central Processing Unit (CPU): More than 100,000 CPU cores and 250 TB of RAM
Graphics Processing Unit (GPU): More than 3.8 million CUDA cores / 240,000 tensor cores of NVIDIA A100 Tensor Core GPU accelerators with a total of 22.4 TB of superfast HBM2 memory
Storage capacity: More than 1 petabyte of user data with high-speed data storage with a speed of 1 TB/s
Applications: Traditional Computational , AI, Big Data

 

DISCOVERER

 

Visual presenting BullSequana XH2000

© Atos

 

Discoverer will be supplied by Atos, based on a BullSequana XH2000 supercomputer and located in Bulgaria.

Sustained performance: 4,44 petaflops
Peak performance: 6 petaflops
Compute partitions: One partition providing 1128 nodes, 4,44 petaflops
Central Processing Unit (CPU): AMD EPYC 7H12 64core, 2.6GHz, 280W (Code name Rome)
Graphics Processing Unit (GPU): No
Storage capacity: petabytes
Applications: Traditional Computational
Other details: Topology - Dragonfly+ with 200Gbps (IB HDR) bandwidth per link

 

DEUCALION

..

© Fujitsu

Deucalion supercomputer will be supplied by Fujitsu and located in Portugal. It will combine a Fujitsu PRIMEHPC (ARM partition) and Atos Bull Sequana (x86 partitions).

Sustained performance: 7,22 petaflops
Peak performance: 10 petaflops
Compute partitions: ARM Partition: 1632 nodes, 3.8 PFLops ; x86 Partition: 500 nodes, 1,62 PFLops ; Accelerated: 33 nodes, 1,72 PFLops
Central Processing Unit (CPU):

A64FX (ARM partition), AMD EPYC (x86 partitions)

Graphics Processing Unit (GPU): NVidia Ampere
Storage capacity:

430 TB High-speed NVMe partition, 10.6 PB high-speed based Parallel File System partition.

Applications: Traditional Computational, AI, Big Data
Other details:

Deucalion will be installed at the Portuguese Foundation for Science and Technology (FCT) Minho Advanced Computing Centre (MACC), in close collaboration with the municipality of Guimarães, in the North of Portugal, as part of a fully sustainable computing infrastructure aiming at promoting new advancements in the digital and green transitions


 

Budget

The EuroHPC Joint Undertaking is jointly funded by its public members with a current budget of around EUR 1.1 billion for the period 2019-2020.

Most of the funding comes from the current EU long-term budget, the Multiannual Financial Framework (MFF) with a contribution of EUR 536 million. This sum is expected to be matched by a similar amount from the participating countries. Private members will also provide additional contributions to the value of over EUR 420 million, through participation in the Joint Undertaking’s activities.

The Joint Undertaking provides financial support in the form of procurement or research and innovation grants to participants following open and competitive calls.

Next steps

HPC is one of the key digital topics where the EU's investment should significantly increase in the next MFF (2021-2027). For this next financial period, the Digital European Programme (DEP), Horizon Europe (H-E) and Connecting Europe Facility-2 (CEF-2) are the main EU funding programmes that could be used to finance the EuroHPC JU.

On 18 September 2020, the European Commission proposed a new Council regulation allowing the EuroHPC JU to continue the development of HPC in Europe for the next decade.

The new regulation aims at replacing the Council Regulation (EU) 2018/1488 establishing the EuroHPC JU. It sets out an ambitious mission to provide Europe with a world-leading hyper-connected supercomputing and quantum computing infrastructure, which will be easily and securely accessible from anywhere in Europe. The new regulation will also enable support to research and innovation activities for new supercomputing technologies, systems applications and products as well as the development of necessary skills to use the infrastructure and form the basis for a world-class HPC ecosystem in Europe.

Benefits of supercomputing

Supercomputing is a critical tool for understanding and responding to complex challenges and transforming them into innovation opportunities.

Benefits for citizens

Supercomputing is starting to play a key role in medicine: for discovering new drugs, developing and targeting medical therapies for the individual needs and conditions of patients experiencing cancer, cardiovascular or Alzheimer’s diseases and rare genetic disorders. Today, supercomputers are actively involved in the quest for treatments for COVID-19 by testing drug candidate molecules or repositioning existing drugs for new diseases. Supercomputing is also crucial to understand the generation and evolution of epidemics and diseases.

Supercomputing is of critical importance to anticipate severe weather conditions: it can provide accurate simulations predicting the evolution of weather patterns, as well as the size and paths of storms and floods. This is key to activate early warning systems to save human lives and reduce damages to our properties and public infrastructures.

Supercomputers are also key to monitor the effects of the climate change. They do so by improving our knowledge of geophysical processes, monitoring earth resource evolution, reducing the environmental footprint of industry and society or supporting sustainable agriculture trough numerical simulations of plant growth.

Supercomputers are also vital for national security, defence and sovereignty, as they are used to increase cybersecurity and in the fight against cyber-criminality, in particular for the protection of critical infrastructures.

Benefits for industry

Supercomputing enables industrial sectors like automotive, aerospace, renewable energy and health to innovate, become more productive and to scale up to higher value products and services.

Supercomputing has a growing impact on industries and businesses by significantly reducing product design and production cycles, accelerating the design of new materials, minimising costs, increasing resource efficiency and shortening and optimising decision processes.

It paves the way to novel industrial applications: from safer and greener vehicles to more efficient photovoltaics, sustainable buildings and optimised turbines for electricity production.

In particular, the use of supercomputing over the cloud will make it easier for SMEs without the financial means to invest in in-house skills to develop and produce better products and services.

Benefits for science

Supercomputing is at the heart of the digital transformation of science. It enables deeper scientific understanding and breakthroughs in nearly every scientific field. 

The applications of supercomputing in science are countless: from fundamental physics (advancing the frontiers of knowledge of matter or exploring the universe) to material sciences (designing new critical components for the pharmaceutical or energy sectors) and earth science (modelling the atmospheric and oceanic phenomena at planetary level).

Many recent breakthroughs would not have been possible without access to the most advanced supercomputers. For example for the Chemistry Nobel Prize winners in 2013, supercomputers were used to develop powerful computing programs and software, to understand and predict complex chemical processes or for the Physics Nobel Prize in 2017 supercomputers helped to make complex calculations to detect hitherto theoretical gravitational waves.