Try   HackMD

Description of HPCFS

Hardware

HPC FS consists of two paritions:

  • haswell from 2016: 20 compute nodes, 1 GPU compute node, 1 login node
  • rome from 2021: 20 compute nodes, 1 login node

Each haswell compute node consists of:

  • 2x Intel Xeon E5-2680V3 12-core processor, 2.5 GHz base clock
  • 64 GB DDR4-2133 ram
  • 250 GB SATA HDD
  • QDR Infiniband
  • 2x 1 Gbps Ethernet

Haswell GPU compute node:

  • 2x Intel Xeon E5-2680V3 12-core processor, 2.5 GHz base clock
  • 256 GB DDR4-2133 ram
  • 250 GB SATA HDD
  • QDR Infiniband
  • 10 Gbps Ethernet
  • 3x NVIDIA K80

Haswell login node:

  • 2x Intel Xeon E5-2680V3 12-core processor, 2.5 GHz base clock
  • 256 GB DDR4-2133 ram
  • 2x 1TB SATA SSD
  • QDR Infiniband
  • 10 Gbps Ethernet
  • NVIDIA K40

Rome compute nodes:

  • 2x AMD EPYC 7402 24-core processor, 2.8 GHz base clock
  • 128 GB DDR4-3200 ram
  • 1 TB NVMe SSD
  • HDR100 Infiniband
  • 2x 1 Gbps Ethernet

Rome login node:

  • 2x AMD EPYC 7302 16-core processor, 3 GHz base clock
  • 512 GB DDR4-3200 ram
  • 2x 1 TB NVMe SSD
  • HDR100 Infiniband
  • 2x 10 Gbps Ethernet
  • NVIDIA A100

HPCFS contains also the following storage systems:

  1. Lustre filesystem:
  • 4x 24 TB Object Storage Targets
  • 96 TB total capacity
  1. Network filesystem:
  • 191 TB total capacity

Software

The operating system at HPCFS is Linux, as is the practice on almost all more powerful HPC systems. Due to the compatibility reasons, the CentOS distribution is selected.

On HPCFS we have installed the following software:

  • Ansys Multiphysics
  • Ansys CFX, Fluent, Maxwell, HFSS
  • OpenFOAM CFD + extend
  • VisIt in ParaView postprocesor
  • Intel F90, CC
  • Octave, R, Mathematica, Matlab
  • OpenMP, OpenMPI, HPMPI, IntelMPI
  • ATLAS, BLAS, BLACS, FFTW, GOTO, MUMPS, NetCDF, HDF5, Sparsekit, Scalapack
tags: HPCFS, hardware