Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.

...

What are the current Cluster queues?
Additional nodes were added to the cluster along with new queues. These changes streamline performance and application integration support. In addition contributed nodes by faculty are listed below. Use of contributed nodes is authorized by node owner. If you are not authorized to use the contributed node(s) you may use the short_allall6, normal_all all6 or long_all all6 queues at lower priority but subject to preemption and other constraints.

...

Node owner

New queue

Old queue

Comments

Public

admin_public

admin

system usage only

 

express_publicpublic6

express

short term high priority, special needs

 

int_publicpublic6

 

interactive jobs, high priority, 4 hour limit

 

exbatch_publicpublic6

 

batch jobs, high priority,low capacity, 2 week limit, exclusive

 

paralleltest_publicpublic6

paralleltest

short term debugging

 

short_gpu

 

2 hour limit

 

normal_gpu

 

72 hour limit

 

long_gpu

 

28 days limit

 

short_publicpublic6

short

2 hour limit

 

parallel_publicpublic6

parallel

2 week limit 2-64cpus

 

normal_publicpublic6

normal

default queue - 3 days limit 1 cpu

 

long_publicpublic6

long

28 days 1 cpu

 

dregs_publicpublic6

dregs

364 days 1 cpu

Public shared

short_all

 

shared across all nodes, lower priority

 

normal_allall6

 

shared across all nodes, lower priority

 

*long_allall6

 

shared across all nodes, lower priority

ATLAS

atlas_prod

 

Physics Atlas support

 

atlas_analysis

 

Physics Atlas support

Khardon

int_khardon

 

contributed nodes, interactive 4 hour limit

 

express_khardon

 

contributed nodes, 30 minute limit

 

short_khardon

 

contributed nodes, 2 hour limit

 

normal_khardon

 

contributed nodes, 3 day limit

 

long_khardon

 

contributed nodes, 14 day limit

Miller

int_miller

 

contributed nodes, interactive 4 hour limit

 

express_miller

 

contributed nodes, 30 minute limit

 

short_miller

 

contributed nodes, 2 hour limit

 

normal_miller

 

contributed nodes, 3 day limit

 

long_miller

 

contributed nodes, 14 day limit

Abriola

int_abriola

 

contributed nodes, interactive 4 hour limit

 

express_abriola

 

contributed nodes, 30 minute limit

 

short_abriola

 

contributed nodes, 2 hour limit

 

normal_abriola

 

contributed nodes, 3 day limit

 

long_abriola

 

contributed nodes, 1 year limit

Napier

int_napier

 

contributed nodes, interactive 4 hour limit

 

express_napier

 

contributed nodes, 30 minute limit

 

short_napier

 

contributed nodes, 2 hour limit

 

normal_napier

 

contributed nodes, 3 day limit

 

long_napier

 

contributed nodes, 1 year limit

...

And extra details by queue name:

-bash-3.2$ bqueues -l atlasnormal_prodpublic6

What is the default queue?
If you do not specify a queue by name in your bsub arguments, your job goes to the default queue, which is normal_public.

...

-bash-3.2$ bsub -R Mem24 -queue normal_public public6 ./myprogram

I see that there are some nodes with more than 32gig ram, such as 48 and 96 gig. How do I access them in exclusive mode since I need almost all the ram and my expected job duration is just minutes?

-bash-3.2$ bsub -q express_public public6 -x -R Mem48 ./myprogram

I have a program that is threaded. Is it possible to submit a job that will "take over" a node, or at least the 4 cores on a single chip?

...

How does one use a node exclusively?
Currently the only queues that allow exclusive use is the express_publicpublic6 and exbatch_publicpublic6 queues. However, not all jobs are suitable, so please inquire with an email to cluster support and describe what you intend to do.

How does one actually invoke a job exclusively?
LSF bsub command has the -x option. To send your job to a node that has extra memory and runs exclusively for hours.
-bash-3.2$ bsub -q exbatch_public public6 -x -R Mem16 ./myprogram

How does one make use of nodes with /scratch2 storage?
Note that is is disk storage and not ram memory.
Access to this storage is by request. Please make this request via cluster-support@tufts.edu.

If you submit a job with the following, LSF will place a job on nodes with /scratch2 partitions.
For example, to request at least 40gig of storage for a job to run in the long_public queue try:

-bash_3.2$ bsub -q long_public public6 -R "scratch2 > 40000" ./your_jobname

...

I have a job in one queue, but would rather have it in another. How do I migrate the job?
Use the lsf command, bmod. For example:
-bash-3.2$ bmod -q express_public public6 <job_number>

This will migrate your job with <job_number> to the express_public queue or some other queue.

...

If you have ssh on your workstation, try the following:
> ssh clustercluster6.uit.tufts.edu ". /etc/profile.d/lsf.sh && bsub -q queuename ./yourprogram"
where queuename is one of the above queues.

Suppose I want to copy data via scp from my bash script that is running on a compute node to the /scratch/utln storage area of the login node. How do I reference it?

scp filename tunictunic6.uit.tufts.edu:/scratch/utln
Note, your utln username is needed.

...

What is the path to reference from a job on a compute node to the storage on the login node?
/cluster/scratch/tunictunic6/utln/ ....

How do I convert mixed case file names in a directory to lower case?

...