Skip to content

Hyperion system

How to connect to Hyperion


Before connecting to Hyperion, ensure you have:

  1. An active account on Hyperion.
  2. An SSH client installed on your local machine. UNIX-based systems, including macOS and Linux, typically come pre-installed with an SSH client. For Windows users, software like PuTTY or MobaXTerm are some recommended options.

SSH Login

To establish a connection to Hyperion, use the following command:

$ ssh <username>

Replace <username> with your actual Hyperion username.

You can also establish direct connection with the login nodes:

$ ssh
$ ssh

Upon successful connection, users will be greeted with a shell on one of Hyperion's login nodes. Here, you can manage your files, compile code, or prepare batch submission scripts.

Security Recommendations

  1. Public Key Authentication: While password-based logins are simple, using SSH key pairs for authentication is much more secure. Implementing public key authentication enhances security by making unauthorized access significantly more challenging.

  2. Monitor Activity: Regularly review the list of active sessions and last logins to your Hyperion account. If you detect any unfamiliar activity, report it to our support team immediately.


Compute Node Range
Proccessor / Total cores
Total Nodes
Public Access
hyperion-[001-007], [023-029], [045-051], [067-073], [089-095], [111-117], [133-139], [155-161], [177-181], [206] Intel Xeon Gold 6342 (Icelake) - 48 cores 256 GB - 60 Yes
hyperion-[008], [030], [052], [074], [096], [118], [140], [162], [182] Intel Xeon Gold 6248R (Cascadelake) - 48 cores 96 GB 2x NVIDIA RTX 3090 24GB 9 Yes
hyperion-[009-022], [031-044], [053-066], [075-088], [097-110], [119-132], [141-154], [163-175] Intel Xeon Gold 6248R (Cascadelake) - 48 cores 96 GB - 111 Yes
hyperion-[176], [180-181], [183-205], [207] Intel Xeon Gold 6248R (Cascadelake) - 48 cores 192 GB - 27 No
hyperion-[208-224],[234-251],[263] Intel Xeon Platinum 8362 (Icelake) - 64 cores 2 TB - 36 Yes
hyperion-[225-233] Intel Xeon Platinum 8368 (Icelake) - 76 cores 2 TB - 9 No
hyperion-252 Intel Xeon Gold 6348 (Icelake) - 56 cores 1 TB 8x NVIDIA A100 PCIe 80GB 1 Yes
hyperion-253 Intel Xeon Gold 6348 (Icelake) - 56 cores 1 TB 8x NVIDIA A100 SXM4 80GB 1 Yes
hyperion-[254,257] Intel Xeon Platinum 8358 (Icelake) - 64 cores 2 TB 8x NVIDIA A100 PCIe 80GB 2 Yes
hyperion-[255-256] AMD EPYC 75F3 (Zen 3) - 64 cores 1 TB 8x NVIDIA A100 SXM4 80GB 2 No
hyperion-[258-259] Intel Xeon Platinum 8362 (Icelake) - 64 cores 4 TB - 2 No
hyperion-[260-261] Intel Xeon Gold 6348 (Icelake) - 56 cores 1.5 TB - 2 No
hyperion-262 Intel Xeon Platinum 8362 (Icelake) - 64 cores 2 TB 1x NVIDIA RTX A6000 48GB 1 Yes
hyperion-263 Intel Xeon Platinum 8362 (Icelake) - 64 cores 2 TB - 1 Yes
hyperion-264 Intel Xeon Platinum 8362 (Icelake) - 64 cores 2 TB - 1 No
hyperion-[265-280] Intel Xeon Platinum 8362 (Icelake) - 64 cores 2 TB - 16 No
hyperion-281 Intel Xeon Platinum 8362 (Icelake) - 32 cores 512 GB - 1 No

The cluster currently integrates nodes with two distinct Intel Xeon microarchitectures: Icelake and Cascadelake. When compiling software, it's imperative to consider the specific microarchitecture to optimize performance.

For interconnection, Hyperion employs Infiniband HDR technology. This provides a bandwidth of up to 200 Gb/s per direction and ensures low latency in data communication between nodes.

Filesystems and IO

Mount point
Home directories /home 50 GB 56 TB storage, dotfiles, config files No
scratch /scratch 1.5 TB 611 TB running jobs No
lscratch /lscratch None - running single node jobs No
data /data 3 TB 917 TB storage No


/data filesystem is mounted as read-only on the compute nodes

Login Nodes

  • Hyperion has 2 login nodes: and
  • Each node has two sockets populated with a 48 core Intel Xeon Platinum 8362 each.
  • Each node has 256 GB of RAM.


Remember that login nodes should only be used to do small tasks or compilation and not to run interactive jobs.

In case a multi-process execution or high memory demanding process is detected on a login node, all the user processes will be terminated, and the user will be banned from the cluster until contacting with

Cluster Architecture Considerations

Hyperion is a heterogeneous cluster, composed of nodes with various microarchitectures including Cascadelake, Icelake, and potentially future Sapphire Rapids nodes.

Compiling Your Code

When compiling your code, it's essential to target the specific microarchitecture of the node you intend to run on. For reliable results and performance, compile and execute programs on nodes of the same microarchitecture. For instance, if you're compiling your code for a Cascadelake node, you should also run it on a Cascadelake node. Using binaries compiled for one microarchitecture on a different microarchitecture may lead to unpredictable behaviors, performance issues, or even application crashes.


All software that we, the staff of the supercomputing center, compile system-wide in a centralized manner is optimized and compiled for both Cascadelake and Icelake microarchitectures. Users can find the same environment modules available, irrespective of the microarchitecture of the node they are working on. This approach ensures compatibility and performance across different node types within Hyperion, allowing users to work seamlessly across various microarchitectures.


Compiling your code

Intel compilers are recommended for building your applications on Hyperion. There is no system default modulefile that takes care of this. Use the module avail command to see what versions are available and load an Intel compiler module before compiling. For example:

$ module load intel/2022a

Notice that when a compiler module is loaded, some environment variables are set or modified to add the paths to certain commands, include files, or libraries, to your environment. This helps to simplify the way you do your work.

As an alternative, Hyperion also offers a collection of open source tools such as compilers or scientific libraries. Use module avail command to see versions available. For example:

module avail intel 
module avail FFTW

To learn more about compilers and scientific libraries checkout Environment Modules and the Compilers sections.