Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.

...

  • Compute nodes are comprised of Intel based Cicso and IBM servers
  • slurm replaces LSF scheduling software
  • ssh logins to compute nodes is restricted
  • cross mounts of local disk storage is removed
  • slurm interactive partition limited to 2 nodes for increased performance and quality of service
  • New login and file transfer nodes for better service

Every 3 years, the Tufts high-performance computing cluster environment is partially refreshed to offer increased capacity.  We did so in 2008 and 2011, and again in 2014.  We have had 4 generations of IBM hardware and we will soon welcome a new generation of Cisco UCS hardware.   Our oldest IBM hardware has been retired and  our newest Cisco hardware has been installed.  We will slowly  integrate current cluster IBM hardware into the new Cisco environment during the  Fall 2014.  Concurrent access to old and new cluster will be available.  This will allow for a smooth smoother and non-disruptive fall transition for users and testing. 

By  late Dec. 2014 the  transition to the new cluster production environment will provide a total of 2680 cores (w/ 1680 IBM and 1000 Cisco cores) as compared to 2110 cores during 2013-2014 academic year .  Due to a significantly different and newer architecture, the 1000 Cisco cores are a lot more powerful.  Network interactivity between nodes is also improved.    In practical terms this refresh represents the largest increase in compute power for the Tufts HPC community.

As part of this hardware refresh, we will also be replacing our LSF scheduler with slurm on the new cluster.  We made this decision as did many other high-performance supercomputing centers worldwide. The benefit to users is build an experience built upon a  common understanding and implementation strategy for scaling HPC jobs.

Slurm replaces LSF

See wiki page aboutslurm for  lsf users. 


User facing access nodes to cluster services

There are two new nodes for interfacing the new cluster. 

  • login.cluster.tufts.edu
  • xfer.cluster.tufts.edu

The login node functions in the same fashion as the old cluster,  cluster6.uit.tufts.edu.  This includes compiling, slurm job submissions, editing, etc.   However the need to transfer data into or out of the cluster is separated from the login node.  A second node, xfer.cluster.tufts.edu is provided for file transfers.    This insures a better quality of service to login node users and minimizes storage related logistics.  Access to xfer.cluster.tufts.edu is via scp, sftp, rsync, and any desktop file transfer programs such as WinScp, Filezilla, etc.  Also one may ssh into the node for initiating transfers as needed from either the headnode, login.cluster.tufts.edu, or desktop.   Node xfer.cluster.tufts.edu is a file transfer only service and not another headnode and access to slurm, compilers,etc.  is defeated.

 

 

 

 

Legacy Cluster

...

Information  for cluster6.uit.tufts.edu

May 2014 Announcement

Every 3 years, the Tufts high-performance computing cluster environment is partially refreshed to offer increased capacity. We did so in 2008 and 2011, and we have just completed our competitive bid for our 2014 refresh purchase. As you may know, we currently have 4 generations of IBM hardware and we are proud to announce that we will soon welcome a new generation of Cisco UCS hardware.

...