ScienceCloud currently provides the compute instances and storage with the following grand totals:
|nodes||virtual CPUs||total RAM|
Nodes are equipped with NVIDIA Tesla T4 and P4 GPUs.
|nodes||GPUs||virtual CPUs||total RAM|
|type||raw capacity||usable capacity|
|Block storage||7.0 PB||2.3 PB|
|Object storage||1.7 PB||0.8 PB with replica-2 |
(or 1.2 PB with ec104)
Every compute node has a non-blocking, redundant, 10gbps link to the internal network. This network is used to access the underlying storage infrastructure and provide network connectivity to the virtual machines.
The uplink to the University network is a redundant 20gbps link.
Current ScienceCloud hypervisors¶
|No. nodes||CPU type @ base clock||Base clock||Architecture||CPU/vCPU||RAM per node||Accelerators|
|188||2x Intel(R) Xeon(R) E5-2640 v3||2.60 GHz||Haswell||16/32||128/256 GB|
|100||2x Intel(R) Xeon(R) E5-2680 v4||2.40 GHz||Broadwell||28/56||256 GB|
|96||2x Intel(R) Xeon(R) Gold 6126||2.60 GHz||Skylake||24/48||384 GB|
|2||2x Intel(R) Xeon(R) Gold 6126||2.60 GHz||Skylake||24/48||384 GB||2x nVidia P4|
|64||2x AMD EPYC 7702||2.0 GHz||Rome||64/64||512 GB|
|10||2x AMD EPYC 7702||2.0 GHz||Rome||64/64||512 GB||4 x nVidia T4|
Both the instance root disk's data and any additionally attached volumes' data are stored on our CEPH cluster and are thus accessed via a redundant 20 or 50 gigabit network link.
To reduce fragmentation flavors are statically mapped to a specific hypervisor type. More specifically:
|Flavor type||Hypervisor CPU|
Xeon E5-2640 v3
Xeon E5-2680 v4
AMD EPYC 7702
Xeon Gold 6126
There are a few cases in which it is desirable to have instances backed by the same generation of hardware (i.e. custom compiled software optimized with specific flags).
This can be accomplished on request by enabling specific flavors that are bound to a specific CPU generation. You should keep in mind that the binding solution will be more prone to issues during hardware maintenance as the instances cannot be moved easily from one host to another.