Difference between revisions of "Teach"
(Created page with "{{Infobox Computer |image=center|300px|thumb |name=Teach Cluster |installed=(orig Feb 2013), Oct 2018 |operatingsystem= Linux (Centos 7.4...") |
|||
Line 23: | Line 23: | ||
From '''<tt>teach01</tt>''' you can compile, do short tests, and submit your jobs to the queue. | From '''<tt>teach01</tt>''' you can compile, do short tests, and submit your jobs to the queue. | ||
− | + | == Software Modules == | |
+ | <pre> | ||
+ | module avail | ||
+ | </pre> | ||
− | |||
+ | == Submit a Job == | ||
+ | Teach uses SLURM as its job scheduler. More-advanced details of how to interact with the scheduler can be found on the [[Slurm | Slurm page]]. | ||
− | + | You submit jobs from a login node by passing a script to the sbatch command: | |
+ | teach01:scratch$ sbatch jobscript.sh | ||
+ | This puts the job in the queue. It will run on the compute nodes in due course. | ||
− | + | In most cases, you will want to submit from your $SCRATCH directory, so that the output of your compute job can be written out (as mentioned above, $HOME is read-only on the compute nodes). | |
− |
Revision as of 14:13, 30 October 2018
Teach Cluster | |
---|---|
Installed | (orig Feb 2013), Oct 2018 |
Operating System | Linux (Centos 7.4) |
Number of Nodes | 42 |
Interconnect | Infiniband (QDR) |
Ram/Node | 64 Gb |
Cores/Node | 16 ( |
Login/Devel Node | teach01 (from teach.scinet) |
Vendor Compilers | icc/gcc |
Queue Submission | slurm |
Specifications
The Teach Cluster consists of 42 repurposed x86_64 nodes each with two octal core Intel Xeon (Sandybridge) E5-2650 2.0GHz CPUs with 64GB of RAM per node. The nodes are interconnected with 2.6:1 blocking QDR Infiniband for MPI communications and disk I/O to the SciNet Niagara filesystems. In total this cluster contains 672 x86_64 cores.
Login/Devel Node
Login via ssh with your scinet account to teach.scinet.utoronto.ca, which will bring directly to teach01 the gateway/devel node for this cluster. From teach01 you can compile, do short tests, and submit your jobs to the queue.
Software Modules
module avail
Submit a Job
Teach uses SLURM as its job scheduler. More-advanced details of how to interact with the scheduler can be found on the Slurm page.
You submit jobs from a login node by passing a script to the sbatch command:
teach01:scratch$ sbatch jobscript.sh
This puts the job in the queue. It will run on the compute nodes in due course.
In most cases, you will want to submit from your $SCRATCH directory, so that the output of your compute job can be written out (as mentioned above, $HOME is read-only on the compute nodes).