On this page:
11.1 Cloud  Lab Utah
11.2 Cloud  Lab Wisconsin
11.3 Cloud  Lab Clemson
11.4 Apt Cluster
11.5 Mass
11.6 One  Lab
2020-07-23 (06836dc)

11 Hardware

CloudLab can allocate experiments on any one of several federated clusters.

CloudLab has the ability to dispatch experiments to several clusters: three that belong to CloudLab itself, plus several more that belong to federated projects.

Additional hardware expansions are planned, and descriptions of them can be found at https://www.cloudlab.us/hardware.php

11.1 CloudLab Utah

The CloudLab cluster at the University of Utah is being built in partnership with HP and Dell. It consists of 28 AMD EPYC Rome servers, 200 Intel Xeon E5 servers, 270 Xeon-D servers, and 315 64-bit ARM servers and 270 Intel Xeon-D severs, for a total of 7,576 cores. The cluster is housed in the University of Utah’s Downtown Data Center in Salt Lake City.

m400

   

315 nodes (64-bit ARM)

CPU

   

Eight 64-bit ARMv8 (Atlas/A57) cores at 2.4 GHz (APM X-GENE)

RAM

   

64GB ECC Memory (8x 8 GB DDR3-1600 SO-DIMMs)

Disk

   

120 GB of flash (SATA3 / M.2, Micron M500)

NIC

   

Dual-port Mellanox ConnectX-3 10 GB NIC (PCIe v3.0, 8 lanes

m510

   

270 nodes (Intel Xeon-D)

CPU

   

Eight-core Intel Xeon D-1548 at 2.0 GHz

RAM

   

64GB ECC Memory (4x 16 GB DDR4-2133 SO-DIMMs)

Disk

   

256 GB NVMe flash storage

NIC

   

Dual-port Mellanox ConnectX-3 10 GB NIC (PCIe v3.0, 8 lanes

There are 45 nodes in a chassis, and this cluster consists of thirteen chassis. Each chassis has two 45XGc switches; each node is connected to both switches, and each chassis switch has four 40Gbps uplinks, for a total of 320Gbps of uplink capacity from each chassis. One switch is used for control traffic, connecting to the Internet, etc. The other is used to build experiment topologies, and should be used for most experimental purposes.

All chassis are interconnected through a large HP FlexFabric 12910 switch which has full bisection bandwidth internally.

We have plans to enable some users to allocate entire chassis; when allocated in this mode, it will be possible to have complete administrator control over the switches in addition to the nodes.

In phase two we added 50 Apollo R2200 chassis each with four HPE ProLiant XL170r server modules. Each server has 10 cores for a total of 2000 cores.

xl170

   

200 nodes (Intel Broadwell, 10 core, 1 disk)

CPU

   

Ten-core Intel E5-2640v4 at 2.4 GHz

RAM

   

64GB ECC Memory (4x 16 GB DDR4-2400 DIMMs)

Disk

   

Intel DC S3520 480 GB 6G SATA SSD

NIC

   

Two Dual-port Mellanox ConnectX-4 25 GB NIC (PCIe v3.0, 8 lanes

Each server is connected via a 10Gbps control link (Dell switches) and a 25Gbps experimental link to Mellanox 2410 switches in groups of 40 servers. Each of the five groups’ experimental switches are connected to a Mellanox 2700 spine switch at 5x100Gbps. That switch in turn interconnects with the rest of the Utah CloudLab cluster via 6x40Gbps uplinks to the HP FlexFabric 12910 switch.

A unique feature of the phase two nodes is the addition of eight ONIE bootable "user allocatable" switches that can run a variety of Open Network OSes: six Dell S4048-ONs and two Mellanox MSN2410-BB2Fs. These switches and all 200 nodes are connected to two NetScout 3903 layer-1 switches, allowing flexible combinations of nodes and switches in an experiment.

For phase two we also added 28 Dell AMD EPYC-based servers with dual 100Gb Ethernet ports.

d6515

   

28 nodes (AMD EPYC Rome, 32 core, 2 disk, 100Gb Ethernet)

CPU

   

32-core AMD 7452 at 2.35GHz

RAM

   

128GB ECC Memory (8x 16 GB 3200MT/s RDIMMs)

Disk

   

Two 480 GB 6G SATA SSD

NIC

   

Dual-port Mellanox ConnectX-5 100 GB NIC (PCIe v4.0)

NIC

   

Dual-port Broadcom 57414 25 GB NIC

Each server is connected via a 25Gbps control link (Dell S5224F switch), 2 x 100Gbs experiment links (Dell Z9264F-ON switch), and a 25Gbps experiment link (Dell S5248F-ON switch). The experiment switches are connected to the "phase two" Mellanox 2700 spine switch at 4x100Gbps (Z9264F) and 2x100Gbps (S5248F).

We hope to make a second server, also with 45TB, available soon.

The Utah Cloudlab cluster includes a storage server for remote datasets. The server currently has 45TB available for allocation.

11.2 CloudLab Wisconsin

The CloudLab cluster at the University of Wisconsin is built in partnership with Cisco, Seagate, and HP. The cluster, which is in Madison, Wisconsin, has 270 servers with a total of 5,000 cores connected in a CLOS topology with full bisection bandwidth. It has 1,070 TB of storage, including SSDs on every node.

More technical details can be found at https://www.cloudlab.us/hardware.php#wisconsin

c220g1

   

90 nodes (Haswell, 16 core, 3 disks)

CPU

   

Two Intel E5-2630 v3 8-core CPUs at 2.40 GHz (Haswell w/ EM64T)

RAM

   

128GB ECC Memory (8x 16 GB DDR4 1866 MHz dual rank RDIMMs)

Disk

   

Two 1.2 TB 10K RPM 6G SAS SFF HDDs

Disk

   

One Intel DC S3500 480 GB 6G SATA SSDs

NIC

   

Dual-port Intel X520-DA2 10Gb NIC (PCIe v3.0, 8 lanes)

NIC

   

Onboard Intel i350 1Gb

c240g1

   

10 nodes (Haswell, 16 core, 14 disks)

CPU

   

Two Intel E5-2630 v3 8-core CPUs at 2.40 GHz (Haswell w/ EM64T)

RAM

   

128GB ECC Memory (8x 16 GB DDR4 1866 MHz dual rank RDIMMs)

Disk

   

Two Intel DC S3500 480 GB 6G SATA SSDs

Disk

   

Twelve 3 TB 3.5" HDDs donated by Seagate

NIC

   

Dual-port Intel X520-DA2 10Gb NIC (PCIe v3.0, 8 lanes)

NIC

   

Onboard Intel i350 1Gb

c220g2

   

163 nodes (Haswell, 20 core, 3 disks)

CPU

   

Two Intel E5-2660 v3 10-core CPUs at 2.60 GHz (Haswell EP)

RAM

   

160GB ECC Memory (10x 16 GB DDR4 2133 MHz dual rank RDIMMs)

Disk

   

One Intel DC S3500 480 GB 6G SATA SSDs

Disk

   

Two 1.2 TB 10K RPM 6G SAS SFF HDDs

NIC

   

Dual-port Intel X520 10Gb NIC (PCIe v3.0, 8 lanes

NIC

   

Onboard Intel i350 1Gb

c240g2

   

4 nodes (Haswell, 20 core, 8 disks)

CPU

   

Two Intel E5-2660 v3 10-core CPUs at 2.60 GHz (Haswell EP)

RAM

   

160GB ECC Memory (10x 16 GB DDR4 2133 MHz dual rank RDIMMs)

Disk

   

Two Intel DC S3500 480 GB 6G SATA SSDs

Disk

   

Two 1TB HDDs

Disk

   

Four 3TB HDDs

NIC

   

Dual-port Intel X520 10Gb NIC (PCIe v3.0, 8 lanes

NIC

   

Onboard Intel i350 1Gb

All nodes are connected to two networks:

The experiment network at Wisconsin is transitioning to using HP switches in order to provide OpenFlow 1.3 support.

Phase II added 260 new nodes, 36 with one or more GPUs:

c220g5

   

224 nodes (Intel Skylake, 20 core, 2 disks)

CPU

   

Two Intel Xeon Silver 4114 10-core CPUs at 2.20 GHz

RAM

   

192GB ECC DDR4-2666 Memory

Disk

   

One 1 TB 7200 RPM 6G SAS HDs

Disk

   

One Intel DC S3500 480 GB 6G SATA SSD

NIC

   

Dual-port Intel X520-DA2 10Gb NIC (PCIe v3.0, 8 lanes)

NIC

   

Onboard Intel i350 1Gb

c240g5

   

32 nodes (Intel Skylake, 20 core, 2 disks, GPU)

CPU

   

Two Intel Xeon Silver 4114 10-core CPUs at 2.20 GHz

RAM

   

192GB ECC DDR4-2666 Memory

Disk

   

One 1 TB 7200 RPM 6G SAS HDs

Disk

   

One Intel DC S3500 480 GB 6G SATA SSD

GPU

   

One NVIDIA 12GB PCI P100 GPU

NIC

   

Dual-port Intel X520-DA2 10Gb NIC (PCIe v3.0, 8 lanes)

NIC

   

Onboard Intel i350 1Gb

c4130

   

4 nodes (Intel Broadwell, 16 core, 2 disks, 4 GPUs)

CPU

   

Two Intel Xeon E5-2667 8-core CPUs at 3.20 GHz

RAM

   

128GB ECC Memory

Disk

   

Two 960 GB 6G SATA SSD

GPU

   

Four NVIDIA 16GB Tesla V100 SMX2 GPUs

The Wisconsin Cloudlab cluster includes a storage server for remote datasets. The server currently has 30TB available for allocation.

11.3 CloudLab Clemson

The CloudLab cluster at Clemson University has been built partnership with Dell. The cluster so far has 260 servers with a total of 6,736 cores, 1,272TB of disk space, and 73TB of RAM. All nodes have 10GB Ethernet and most have QDR Infiniband. It is located in Clemson, South Carolina.

More technical details can be found at https://www.cloudlab.us/hardware.php#clemson

c8220

   

96 nodes (Ivy Bridge, 20 core)

CPU

   

Two Intel E5-2660 v2 10-core CPUs at 2.20 GHz (Ivy Bridge)

RAM

   

256GB ECC Memory (16x 16 GB DDR4 1600MT/s dual rank RDIMMs

Disk

   

Two 1 TB 7.2K RPM 3G SATA HDDs

NIC

   

Dual-port Intel 10Gbe NIC (PCIe v3.0, 8 lanes

NIC

   

Qlogic QLE 7340 40 Gb/s Infiniband HCA (PCIe v3.0, 8 lanes)

c8220x

   

4 nodes (Ivy Bridge, 20 core, 20 disks)

CPU

   

Two Intel E5-2660 v2 10-core CPUs at 2.20 GHz (Ivy Bridge)

RAM

   

256GB ECC Memory (16x 16 GB DDR4 1600MT/s dual rank RDIMMs

Disk

   

Eight 1 TB 7.2K RPM 3G SATA HDDs

Disk

   

Twelve 4 TB 7.2K RPM 3G SATA HDDs

NIC

   

Dual-port Intel 10Gbe NIC (PCIe v3.0, 8 lanes

NIC

   

Qlogic QLE 7340 40 Gb/s Infiniband HCA (PCIe v3.0, 8 lanes)

c6320

   

84 nodes (Haswell, 28 core)

CPU

   

Two Intel E5-2683 v3 14-core CPUs at 2.00 GHz (Haswell)

RAM

   

256GB ECC Memory

Disk

   

Two 1 TB 7.2K RPM 3G SATA HDDs

NIC

   

Dual-port Intel 10Gbe NIC (X520)

NIC

   

Qlogic QLE 7340 40 Gb/s Infiniband HCA (PCIe v3.0, 8 lanes)

c4130

   

2 nodes (Haswell, 28 core, two GPUs)

CPU

   

Two Intel E5-2680 v3 12-core processors at 2.50 GHz (Haswell)

RAM

   

256GB ECC Memory

Disk

   

Two 1 TB 7.2K RPM 3G SATA HDDs

GPU

   

Two Tesla K40m GPUs

NIC

   

Dual-port Intel 1Gbe NIC (i350)

NIC

   

Dual-port Intel 10Gbe NIC (X710)

NIC

   

Qlogic QLE 7340 40 Gb/s Infiniband HCA (PCIe v3.0, 8 lanes)

There are also two, storage intensive (270TB each!) nodes that should only be used if you need a huge amount of volatile storage. These nodes have only 10GB Ethernet.

dss7500

   

2 nodes (Haswell, 12 core, 270TB disk)

CPU

   

Two Intel E5-2620 v3 6-core CPUs at 2.40 GHz (Haswell)

RAM

   

128GB ECC Memory

Disk

   

Two 120 GB 6Gbps SATA SSDs

Disk

   

45 6 TB 7.2K RPM 6Gbps SATA HDDs

NIC

   

Dual-port Intel 10Gbe NIC (X520)

There are three networks at the Clemson site:

Phase two added 18 Dell C6420 chassis each with four dual-socket Skylake-based servers. Each of the 72 servers has 32 cores for a total of 2304 cores.

c6420

   

72 nodes (Intel Skylake, 32 core, 2 disk)

CPU

   

Two Sixteen-core Intel Xeon Gold 6142 CPUs at 2.6 GHz

RAM

   

384GB ECC DDR4-2666 Memory

Disk

   

Two Seagate 1TB 7200 RPM 6G SATA HDs

NIC

   

Dual-port Intel X710 10Gbe NIC

Each server is connected via a 1Gbps control link (Dell D3048 switches) and a 10Gbps experimental link (Dell S5048 switches).

These Phase II machines do not include Infiniband.

Phase II also added 6 IBM Power System S822LC (8335-GTB) POWER8 servers. These machines are booted using the Linux-based OpenPOWER firmware (OPAL). They can run code in either little- or big-endian modes, but we only provide a little-endian standard system image (‘UBUNTU18-PPC64LE‘).

ibm8335

   

6 nodes (POWER8NVL, 20 core, 256GB RAM, 1 GPU)

CPU

   

Two ten-core (8 threads/core) IBM POWER8NVL CPUs at 2.86 GHz

RAM

   

256GB 1600MHz DDR4 memory

Disk

   

Two Seagate 1TB 7200 RPM 6G SATA HDDs (ST1000NX0313)

NIC

   

One Broadcom NetXtreme II BCM57800 1/10 GbE NIC

GPU

   

Two NVIDIA GP100GL (Tesla P100 SMX2 16GB)

FPGA

   

One ADM-PCIE-KU3 (Xilinx Kintex UltraScale)

You can find more info here.

Phase III added 15 Dell R7525 servers, each with dual-core AMD EPYX processors, two NVIDIA GPUs, and a Mellanox BlueField2 SmartNIC.

r7525

   

15 nodes (AMD EPYC Rome, 64 core, 512GB RAM, 2 x GPU)

CPU

   

Two 32-core AMD 7542 at 2.9GHz

RAM

   

512GB ECC Memory (16x 32 GB 3200MHz DDR4)

Disk

   

One 2TB 7200 RPM 6G SATA HDD

NIC

   

Dual-port Mellanox ConnectX-5 25 Gb NIC (PCIe v4.0)

NIC

   

Dual-port Mellanox BlueField2 100 Gb SmartNIC

GPU

   

Two NVIDIA GV100GL (Tesla V100S PCIe 32GB)

As of July 2020, the BlueField2 NICs are not yet installed.

The nodes have a 1Gb control network connection, one 25Gb experiment connection, and 2 x 100Gb connections via the BlueField2 card.

The Clemson Cloudlab cluster includes a storage server for remote datasets. The server currently has 40TB available for allocation.

11.4 Apt Cluster

The main Apt cluster is housed in the University of Utah’s Downtown Data Center in Salt Lake City, Utah. It contains two classes of nodes:

r320

   

128 nodes (Sandy Bridge, 8 cores)

CPU

   

1x Xeon E5-2450 processor (8 cores, 2.1Ghz)

RAM

   

16GB Memory (4 x 2GB RDIMMs, 1.6Ghz)

Disks

   

4 x 500GB 7.2K SATA Drives (RAID5)

NIC

   

1GbE Dual port embedded NIC (Broadcom)

NIC

   

1 x Mellanox MX354A Dual port FDR CX3 adapter w/1 x QSA adapter

c6220

   

64 nodes (Ivy Bridge, 16 cores)

CPU

   

2 x Xeon E5-2650v2 processors (8 cores each, 2.6Ghz)

RAM

   

64GB Memory (8 x 8GB DDR-3 RDIMMs, 1.86Ghz)

Disks

   

2 x 1TB SATA 3.5” 7.2K rpm hard drives

NIC

   

4 x 1GbE embedded Ethernet Ports (Broadcom)

NIC

   

1 x Intel X520 PCIe Dual port 10Gb Ethernet NIC

NIC

   

1 x Mellanox FDR CX3 Single port mezz card

All nodes are connected to three networks with one interface each:

There is no remote dataset capability at the Apt cluster.

11.5 Mass

UMass and the Mass Open Cloud host a cluster at the Massachusetts Green High Performance Compute Center in Holyoke, Massachusetts.

rs440

   

5 nodes (Skylake, 32 cores)

CPU

   

2 x Xeon Gold 6130 processors (16 cores each, 2.1Ghz)

RAM

   

192GB Memory (12 x 16GB RDIMMs)

Disks

   

1 x 240GB SATA SSD drives

NIC

   

2 x 10GbE embedded Ethernet Ports (Broadcom 57412)

These nodes are connected via two 10Gbps ports to a Dell S4048-ON switch. One port is used for control traffic and connectivity to the public Internet, and the other is used for the experiment network.

rs620

   

11 nodes (Sandy Bridge, 16 or 20 cores)

CPU

   

2 x Xeon processors (8-10 cores each, 2.2Ghz or more)

RAM

   

128-384GB Memory (most have 256GB)

Disks

   

1 x 900GB 10K SAS Drive

NIC

   

1GbE Quad port embedded NIC (Intel)

NIC

   

1 x Solarflare Dual port SFC9120 10G Ethernet NIC

rs630

   

28 nodes (Haswell, 20 cores)

CPU

   

2 x Xeon E5-2660 v3 processors (10 cores each, 2.6Ghz or more)

RAM

   

256GB Memory (16 x 16GB DDR4 DIMMs)

Disks

   

1 x 900GB 10K SAS Drive

NIC

   

1GbE Quad port embedded NIC (Intel)

NIC

   

1 x Solarflare Dual port SFC9120 10G Ethernet NIC

There is some variation within the rs620 and rs630 nodes, primarily with the CPUs.

On these nodes, the control/Internet connections is a 1Gbps port and one of the 10Gbps interfaces on each node is used for the experiment network.

There is currently no remote dataset capability at the UMass cluster.

11.6 OneLab

The OneLab facility at Sorbonne University in Paris hosts a small cluster modeled after part of the Utah hardware, with one chassis of ARM64 servers. In addition to this cluster, which is available to all CloudLab users through the CloudLab interface, OneLab hosts a large number of other epxeriment envirnoments, including clusters, IoT devices, and software defined networks. See the OneLab website for a complete list.

m400

   

45 nodes (64-bit ARM)

CPU

   

Eight 64-bit ARMv8 (Atlas/A57) cores at 2.4 GHz (APM X-GENE)

RAM

   

64GB ECC Memory (8x 8 GB DDR3-1600 SO-DIMMs)

Disk

   

120 GB of flash (SATA3 / M.2, Micron M500)

NIC

   

Dual-port Mellanox ConnectX-3 10 GB NIC (PCIe v3.0, 8 lanes

There is no remote dataset capability at the OneLab cluster.