<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>https://docs.scinet.utoronto.ca/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Rzon</id>
	<title>SciNet Users Documentation - User contributions [en]</title>
	<link rel="self" type="application/atom+xml" href="https://docs.scinet.utoronto.ca/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Rzon"/>
	<link rel="alternate" type="text/html" href="https://docs.scinet.utoronto.ca/index.php/Special:Contributions/Rzon"/>
	<updated>2026-05-08T12:47:05Z</updated>
	<subtitle>User contributions</subtitle>
	<generator>MediaWiki 1.35.12</generator>
	<entry>
		<id>https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7715</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7715"/>
		<updated>2026-05-01T16:56:09Z</updated>

		<summary type="html">&lt;p&gt;Rzon: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;__NOTOC__&lt;br /&gt;
{| style=&amp;quot;border-spacing:10px; width: 95%&amp;quot;&lt;br /&gt;
| style=&amp;quot;padding:1em; padding-top:.1em; border:2px solid #0645ad; background-color:#f6f6f6; border-radius:7px&amp;quot;|&lt;br /&gt;
&lt;br /&gt;
==System Status==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- Use &amp;quot;Up&amp;quot;, &amp;quot;Partial&amp;quot; or &amp;quot;Down&amp;quot;; these are templates. --&amp;gt;&lt;br /&gt;
{|style=&amp;quot;width:100%&amp;quot; &lt;br /&gt;
|{{Up3 | Trillium|https://docs.alliancecan.ca/wiki/Trillium_Quickstart}}&lt;br /&gt;
|{{Up3 | OnDemand|https://docs.alliancecan.ca/wiki/Trillium_Open_OnDemand_Quickstart}}&lt;br /&gt;
|{{Up | Globus |Globus}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Up | HPSS|HPSS}}&lt;br /&gt;
|{{up | Balam|Balam}}&lt;br /&gt;
|{{Up | S4H | S4H}}&lt;br /&gt;
|-&lt;br /&gt;
|{{up | Teach|Teach}}&lt;br /&gt;
|{{Up3 | File system|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Storage}}&lt;br /&gt;
|{{Up3 | External Network|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Logging_in}} &lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Wed Apr 30, 2026, 3:00 pm:''' System have been updated to mitigate known security risks, and are back in service. Note that no actual security breaches were found.&lt;br /&gt;
&lt;br /&gt;
'''Wed Apr 29, 2026, 5:25 pm:''' For security reasons, login access to all systems has been disabled, as have OnDemand Apps.  Compute jobs are still allowed to run.&lt;br /&gt;
 &lt;br /&gt;
'''Thu Apr 23, 2026, 10:00 am:''' The Trillium file system has issues and may be slow on certain nodes. We are still investigating.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--  When removing system status entries, please archive them to: --&amp;gt;&lt;br /&gt;
[[Previous messages]]&lt;br /&gt;
&lt;br /&gt;
{|style=&amp;quot;border-spacing: 10px;width: 100%&amp;quot;&lt;br /&gt;
|valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== QuickStart Guides ==&lt;br /&gt;
* [https://docs.alliancecan.ca/wiki/Trillium_Quickstart Trillium Quickstart]&lt;br /&gt;
* [[HPSS | HPSS archival storage]]&lt;br /&gt;
* [[Teach|Teach cluster]]&lt;br /&gt;
* [[FAQ | FAQ (frequently asked questions)]]&lt;br /&gt;
* [[Acknowledging SciNet]]&lt;br /&gt;
| valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== Tutorials, Manuals, etc. ==&lt;br /&gt;
* [https://education.scinet.utoronto.ca SciNet education material]&lt;br /&gt;
* [https://www.youtube.com/c/SciNetHPCattheUniversityofToronto SciNet's YouTube channel]&lt;br /&gt;
* [[Modules for Mist]] &lt;br /&gt;
* [[Commercial software]]&lt;br /&gt;
* [[Burst Buffer]]&lt;br /&gt;
* [[SSH#SSH Keys|SSH keys]]&lt;br /&gt;
* [[SSH Tunneling]]&lt;br /&gt;
* [[Visualization]]&lt;br /&gt;
* [[Running Serial Jobs on Niagara]]&lt;br /&gt;
* [[Jupyter Hub]]&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Rzon</name></author>
	</entry>
	<entry>
		<id>https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7712</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7712"/>
		<updated>2026-05-01T16:00:23Z</updated>

		<summary type="html">&lt;p&gt;Rzon: /* System Status */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;__NOTOC__&lt;br /&gt;
{| style=&amp;quot;border-spacing:10px; width: 95%&amp;quot;&lt;br /&gt;
| style=&amp;quot;padding:1em; padding-top:.1em; border:2px solid #0645ad; background-color:#f6f6f6; border-radius:7px&amp;quot;|&lt;br /&gt;
&lt;br /&gt;
==System Status==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- Use &amp;quot;Up&amp;quot;, &amp;quot;Partial&amp;quot; or &amp;quot;Down&amp;quot;; these are templates. --&amp;gt;&lt;br /&gt;
{|style=&amp;quot;width:100%&amp;quot; &lt;br /&gt;
|{{Up3 | Trillium|https://docs.alliancecan.ca/wiki/Trillium_Quickstart}}&lt;br /&gt;
|{{Up3 | OnDemand|https://docs.alliancecan.ca/wiki/Trillium_Open_OnDemand_Quickstart}}&lt;br /&gt;
|{{Up | Globus |Globus}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Up | HPSS|HPSS}}&lt;br /&gt;
|{{up | Balam|Balam}}&lt;br /&gt;
|{{Up | S4H | S4H}}&lt;br /&gt;
|-&lt;br /&gt;
|{{up | Teach|Teach}}&lt;br /&gt;
|{{Up3 | File system|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Storage}}&lt;br /&gt;
|{{Up3 | External Network|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Logging_in}} &lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Wed Apr 30, 2026, 3:00 pm:''' Most system have been updated to mitigate known security risks, and are back in service. Note that no actual security breaches were found.&lt;br /&gt;
&lt;br /&gt;
'''Wed Apr 29, 2026, 5:25 pm:''' For security reasons, login access to all systems has been disabled, as have OnDemand Apps.  Compute jobs are still allowed to run.&lt;br /&gt;
 &lt;br /&gt;
'''Thu Apr 23, 2026, 10:00 am:''' The Trillium file system has issues and may be slow on certain nodes. We are still investigating.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--  When removing system status entries, please archive them to: --&amp;gt;&lt;br /&gt;
[[Previous messages]]&lt;br /&gt;
&lt;br /&gt;
{|style=&amp;quot;border-spacing: 10px;width: 100%&amp;quot;&lt;br /&gt;
|valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== QuickStart Guides ==&lt;br /&gt;
* [https://docs.alliancecan.ca/wiki/Trillium_Quickstart Trillium Quickstart]&lt;br /&gt;
* [[HPSS | HPSS archival storage]]&lt;br /&gt;
* [[Teach|Teach cluster]]&lt;br /&gt;
* [[FAQ | FAQ (frequently asked questions)]]&lt;br /&gt;
* [[Acknowledging SciNet]]&lt;br /&gt;
| valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== Tutorials, Manuals, etc. ==&lt;br /&gt;
* [https://education.scinet.utoronto.ca SciNet education material]&lt;br /&gt;
* [https://www.youtube.com/c/SciNetHPCattheUniversityofToronto SciNet's YouTube channel]&lt;br /&gt;
* [[Modules for Mist]] &lt;br /&gt;
* [[Commercial software]]&lt;br /&gt;
* [[Burst Buffer]]&lt;br /&gt;
* [[SSH#SSH Keys|SSH keys]]&lt;br /&gt;
* [[SSH Tunneling]]&lt;br /&gt;
* [[Visualization]]&lt;br /&gt;
* [[Running Serial Jobs on Niagara]]&lt;br /&gt;
* [[Jupyter Hub]]&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Rzon</name></author>
	</entry>
	<entry>
		<id>https://docs.scinet.utoronto.ca/index.php?title=Previous_messages&amp;diff=7709</id>
		<title>Previous messages</title>
		<link rel="alternate" type="text/html" href="https://docs.scinet.utoronto.ca/index.php?title=Previous_messages&amp;diff=7709"/>
		<updated>2026-05-01T15:59:21Z</updated>

		<summary type="html">&lt;p&gt;Rzon: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
'''Tue Apr 14, 2026, 10:30 am:''' The Trillium globus endpoint is working again.&lt;br /&gt;
&lt;br /&gt;
'''Sat Apr 11, 2026, 10:00 pm:''' The Trillium globus endpoint is not operational (it times out). We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''Thu Apr 09, 2026, 10:30 am:''' tri-dm4.scinet.utoronto.ca and robot4.scinet.utoronto.ca are in maintenance. Use 1,2, or 3 instead.&lt;br /&gt;
&lt;br /&gt;
'''Wed Apr 08, 2026, 6:30 pm:''' Software from the CVMFS 'restricted' area is back.&lt;br /&gt;
&lt;br /&gt;
'''Wed Apr 08, 2026, 1:40 pm:''' Software from the CVMFS 'restricted' area is not available on many of the Trillium and Open OnDemand nodes. We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''Mon Apr 06, 2026, 10:00 pm:''' We will have to reschedule the HPSS update. This attempt didn't work as expected&lt;br /&gt;
&lt;br /&gt;
'''Mon Apr 06, 2026, 8:00 pm:''' HPSS scheduled maintenance: update of HPSS to v11.3_u4 and hsi-htar to v11.3_u1 (bug fixes)&lt;br /&gt;
&lt;br /&gt;
'''Wed Mar 25, 2026, 5:00 pm:''' Trillium is operational again.&lt;br /&gt;
&lt;br /&gt;
'''Wed Mar 25, 2026, 9:00 am:''' Teach is operational again.&lt;br /&gt;
&lt;br /&gt;
'''Tue Mar 24, 2026, 8:45 pm:''' Open OnDemand is operational again.&lt;br /&gt;
&lt;br /&gt;
'''Tue Mar 24, 2026, 1:00 pm:''' External connectivity is back. &lt;br /&gt;
&lt;br /&gt;
'''Tue Mar 24, 2026, 12:05 pm:''' External connectivity to the data centre was lost. &lt;br /&gt;
&lt;br /&gt;
'''Tue Mar 24, 2026, 7:00 am:''' Maintenance has started.&lt;br /&gt;
&lt;br /&gt;
'''Mon Mar 16, 2026. 13:30pm''' Recovering.  Almost all systems are up again. Please resubmit your jobs that crashed. &lt;br /&gt;
&lt;br /&gt;
'''Mon Mar 16, 2026. 12:00pm''' Power glitch at the data centre caused compute nodes to go down.&lt;br /&gt;
&lt;br /&gt;
'''Thu Mar 12, 2026, 4:15 pm''' Connection to Trillium are operational again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Mar 12, 2026, 1:00 pm''' We've had some login issues particularly for Trillium-GPU. We're investigating.&lt;br /&gt;
&lt;br /&gt;
'''Downtime Announcement:'''  The winter cooling tower maintenance for the SciNet data centre will take place on March 24 and 25, 2026, starting at 7:00 a.m. on the 24th.  All SciNet systems (Trillium, OnDemand, Balam, S4H, Teach, as well as hosted equipment) will have their compute nodes shut down. Login nodes, file systems, and the HPSS system will remain available, and&lt;br /&gt;
jobs will be held in the queue until maintenance is complete.  Starting 7am on Mar 23, users are encouraged to submit small and short jobs that may be scheduled before the maintenance begins.&lt;br /&gt;
&lt;br /&gt;
'''Fri Feb 20, 2026, 11:35 pm:''' Power glitch, ~480 compute nodes rebooted. Regional power quality has been quite poor lately ([https://www.yorkregion.com/news/road-salt-blamed-for-power-outages/article_1a36d25d-5f97-56ee-a0c7-c49c7b732d38.html 1],&lt;br /&gt;
[https://www.yorkregion.com/news/power-company-executive-responds-to-york-region-outages/article_c4d072e7-2892-5c9c-8deb-ac5e1936779c.html 2]).&lt;br /&gt;
&lt;br /&gt;
'''Thu Feb 19, 2026, 3:00 pm:''' Systems restored. Please report issues to support@scinet.utoronto.ca.&lt;br /&gt;
&lt;br /&gt;
'''Tue Feb 17, 2026, 8:40 am:''' Power outage at the data centre.  Cooling issues have developed as a result.  Major systems (Trillium, S4H) are expected to be down until sometime Thursday. Login nodes and file systems will remain accessible.&lt;br /&gt;
&lt;br /&gt;
'''Mon Feb 16, 2026, 8:40 pm:''' Electricity is unstable in the data centre area due to severe snowfall.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 1:40 pm:''' All services are operational again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 12:00 pm:''' The Trillium and Open OnDemand compute nodes are operational again. We are still working on bringing Balam, Neptune and S4H nodes up again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 10:00 am:''' There was a power glitch at the data centre overnight. The login nodes are accessible but the compute nodes are down.  &lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 16, 2026, 11:00 pm:''' HPSS is back online, and accessible via alliancecan#hpss Globus endpoint. &lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 15, 2026, 10:00 pm:''' HPSS will undergo maintenance on Friday morning, Jan/16/2025, , including alliancecan#hpss Globus endpoint &lt;br /&gt;
&lt;br /&gt;
'''Tue Jan 6, 2026, 10:15 am:''' OnDemand has been fixed and is working again.&lt;br /&gt;
&lt;br /&gt;
'''Mon Jan 5, 2026, 9:00 pm:''' The authentication mechanism of OnDemand is not working.&lt;br /&gt;
&lt;br /&gt;
'''Wed Dec 31, 2025, 12:40 pm:''' We believe the problem has now been resolved.  Please let us know if you still experience login problems or aborted jobs.&lt;br /&gt;
&lt;br /&gt;
'''Tue Dec 30, 2025, 2:10 pm:''' We are experiencing problems with authentication, resulting in failed logins, OOD errors, and aborted jobs (with &amp;quot;prolog error&amp;quot;).  Please bear with us, as we are very short-staffed during the holiday break.  We will post updates here.&lt;br /&gt;
&lt;br /&gt;
'''Tue Dec 3, 2025, 11:30 am:''' Open OnDemand is fully operational again.&lt;br /&gt;
&lt;br /&gt;
'''Sat Nov 29, 2025, 00:40 am:''' There has been a problem with the water chiller. Some systems are offline.&lt;br /&gt;
&lt;br /&gt;
'''Wed Nov 5, 2025, 12:55 pm:''' Balam is back online.&lt;br /&gt;
&lt;br /&gt;
'''Wed Nov 5, 2025, 10:00 am:''' Open OnDemand is back online.&lt;br /&gt;
&lt;br /&gt;
'''Tue Nov 4, 2025, 11:00 pm:''' Most of the work is done, data movers, Globus, and HPSS are back online. Remaining services will be worked on tomorrow.&lt;br /&gt;
&lt;br /&gt;
'''Tue Nov 4, 2025, 8:30 am:''' Scheduled network maintenance. Trillium cluster is *not* affected.&lt;br /&gt;
&lt;br /&gt;
'''Tue Oct 21, 2025, 17:30 am:''' Balam maintenance finished.&lt;br /&gt;
&lt;br /&gt;
'''Tue Oct 21, 2025, 7:00 am:''' Balam maintenance day.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 15, 2025, 3:55 pm:''' Trillium inbound connections through trillium.alliancecan.ca or trillium.scinet.utoronto.ca are working again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 15, 2025, 3:05 pm:''' Trillium is experiencing external network issues for both incoming traffic. Please try: ssh USERNAME@tri-login01.scinet.utoronto.ca in the meantime.&lt;br /&gt;
 &lt;br /&gt;
'''Thu Oct 06, 2025, 8:00 pm:''' HPSS is fully functional. You may submit archive jobs from trillium login nodes, datamovers and robots.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 03, 2025, 6:30 pm:''' HPSS is back online, and already accessible via alliancecan#hpss Globus endpoint. Directory tree now follows the other Alliance clusters. We're still working on job submission via Slurm&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 01, 2025, 0:00 am:''' Niagara compute nodes are now unavailable for regular users. The login nodes will remain available for a while to allow a few last data transfers, although transfers from the Niagara file systems to Trillium are best done on nia-dm1.scinet.utoronto.ca.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 01, 2025, 9:30 am:''' HPSS is down for scheduled maintenance, including alliancecan#hpss Globus endpoint&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
'''Thu Sep 18, 2025, 11:30 am:''' Open OnDemand is fully functional again.&lt;br /&gt;
&lt;br /&gt;
'''Wed Sep 17, 2025, 6:00 pm:''' Niagara is back up as well (including its Globus endpoint).  We are still working on the other systems.&lt;br /&gt;
&lt;br /&gt;
'''Wed Sep 17, 2025, 1:40 pm:''' Trillium is back up (except for its Globus endpoint).  We are working on the other systems still.&lt;br /&gt;
&lt;br /&gt;
'''Tue Sep 16, 2025, 5:45 pm:''' Unfortunately, we cannot bring all systems up yet because we are waiting for a spare part for the cooling system that will be brought tomorrow.  In the meantime, we have managed to keep the Trillium login nodes up, but not other systems.&lt;br /&gt;
&lt;br /&gt;
'''Tue Sep 16, 2025, from 7:00 am to 5:00 pm (EDT):''' The SciNet datacentre will undergo maintenance of several critical parts of the centre.  This will require a full shutdown of all SciNet systems (Trillium, Niagara, Mist, HPSS, Rouge, Teach, as well as hosted equipment). This will also be the time that the Mist cluster gets decommissioned. &lt;br /&gt;
&lt;br /&gt;
'''Fri Sep 12 22:03:17 EDT 2025:''' HPSS software and OS upgrades are finished.&lt;br /&gt;
&lt;br /&gt;
'''Tue Sep  9 17:05:38 EDT 2025:''' Starting tomorrow, Sep/10, and for the following 3 days HPSS will be down for software and OS upgrades. We will strive to finish sooner, at which time we will make the system available to users again.&lt;br /&gt;
&lt;br /&gt;
===Mist/Niagara Decommissioning Schedule===&lt;br /&gt;
&lt;br /&gt;
'''September 4, 2025'''&lt;br /&gt;
* Niagara reduced to 863 compute nodes.&lt;br /&gt;
&lt;br /&gt;
'''September 9, 2025'''&lt;br /&gt;
* Niagara's Open OnDemand decommissioned.&lt;br /&gt;
* Brief data centre connection outage at 9 AM EDT&lt;br /&gt;
* Niagara reduced to 647 compute nodes at end of day.&lt;br /&gt;
&lt;br /&gt;
'''September 11, 2025'''&lt;br /&gt;
* Trillium Open OnDemand goes live.&lt;br /&gt;
&lt;br /&gt;
'''September 16, 2025'''&lt;br /&gt;
* '''Full-day data centre maintenance'''&lt;br /&gt;
* Niagara reduced to 431 compute nodes.&lt;br /&gt;
* Mist decommissioned.&lt;br /&gt;
&lt;br /&gt;
'''September 24, 2025'''&lt;br /&gt;
* Niagara reduced to 215 compute nodes.&lt;br /&gt;
&lt;br /&gt;
'''September 30, 2025'''&lt;br /&gt;
* Niagara decommissioned.&lt;br /&gt;
&lt;br /&gt;
'''August 25, 2025, 9:50 EDT:''' Open Ondemand is fully operational again.&lt;br /&gt;
&lt;br /&gt;
'''August 22, 2025, 3:15 PM EDT:''' Open Ondemand has issues launching new interactive apps. We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''August 20, 2025, 10:00 AM EDT:''' The GPU scheduler on Trillium is fully operational again.&lt;br /&gt;
&lt;br /&gt;
'''August 19, 2025, 5:00 PM EDT:''' The GPU scheduler on Trillium has trouble scheduling multi-GPU jobs.  We're investigating the issue.&lt;br /&gt;
&lt;br /&gt;
'''August 7, 2025:''' CVMFS issues are resolved.&lt;br /&gt;
&lt;br /&gt;
'''August 6, 2025:''' We are seeing intermittent issues with the software on CVMFS on Niagara. We're investigating.&lt;br /&gt;
&lt;br /&gt;
'''July 31, 2025, 4:00 PM EDT - 5:00 PM EDT:''' As announced, all systems connected to the Niagara file system (Mist, Niagara, HPSS, Balam, and Rouge) will be paused and inaccessible for one hour to start the transfer of files from the Niagara file system to the Trillium file system. &lt;br /&gt;
&lt;br /&gt;
'''January 6, 2025:''' As part of the installation of the new computing cluster Trillium, there is now a permanent reduction in computing capacity of Niagara to 50% and of Mist to 35%.&lt;br /&gt;
&lt;br /&gt;
'''July 9, 2025:''' The [[Teach]] cluster will be unavailable for the day for network maintenance.&lt;br /&gt;
&lt;br /&gt;
'''July 4, 2025:''' Open OnDemand is back up.&lt;br /&gt;
&lt;br /&gt;
'''July 4, 2025:''' Open OnDemand is down. We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''June 25, 2025, 7:15 PM EDT:''' The [[Teach]] cluster's scheduler is up again.&lt;br /&gt;
&lt;br /&gt;
'''June 25, 2025, 4:30 PM EDT:''' The [[Teach]] cluster's scheduler is down. We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''April 30, 2025, 9:30 AM EDT:''' The [[Teach]] cluster is available again.&lt;br /&gt;
&lt;br /&gt;
'''April 30, 2025:''' The [[Teach]] cluster will be unavailable from 8:00 am to about 12:00 noon for file system maintenance.&lt;br /&gt;
&lt;br /&gt;
'''April 1, 2025:''' The Jupyter Hub has been replaced by SciNet's [[Open OnDemand Quickstart|Open OnDemand service]].&lt;br /&gt;
&lt;br /&gt;
'''March 1, 2025:''' As of March 1st scratch purging is suspended until after Trillium comes online.&lt;br /&gt;
&lt;br /&gt;
'''April 15, 2025 12:40 pm EDT: '''Balam login node is accessible again.&lt;br /&gt;
&lt;br /&gt;
'''April 15, 2025 12:10 pm EDT: '''Balam login node is under maintenance and temporarily inaccessible to users.&lt;br /&gt;
&lt;br /&gt;
'''April 9, 2025 9PM:''' HPSS is back online.&lt;br /&gt;
&lt;br /&gt;
'''April 8, 2025 9PM:''' HPSS is being reserved for OS updates on April 9 (Wednesday).&lt;br /&gt;
&lt;br /&gt;
'''March 31, 2025 3:20 pm EDT:''' Mist login node is accessible again.&lt;br /&gt;
&lt;br /&gt;
'''March 31, 2025 2:45 pm EDT:''' Mist login node is under maintenance and temporarily inaccessible to users.&lt;br /&gt;
&lt;br /&gt;
'''March 28, 2025 3:00 pm - 4:00 pm EDT:''' A short maintenance was needed for the Teach compute nodes; you might have experienced some job scheduling delays on that cluster. &lt;br /&gt;
&lt;br /&gt;
'''March 20, 2025 10:30 am EDT:''' Teach compute nodes are back. &lt;br /&gt;
&lt;br /&gt;
'''March 19, 2025 11:00 pm EDT:''' Teach compute nodes are down again. &lt;br /&gt;
&lt;br /&gt;
'''March 19, 2025 5:15pm EDT:''' Maintenance of the cooling system was performed successfully. The cluster is back on line&lt;br /&gt;
&lt;br /&gt;
'''March 19, 2025 8:00 am - 5:00 pm EDT:''' Maintenance of the cooling system as well as preparations for the Trillium cluster will require a shutdown of the compute nodes of all SciNet systems (Niagara, Mist, Rouge, Balam, Teach, as well as hosted equipment). The login nodes, file systems and the HPSS system will remain available. The scheduler will hold jobs that are submitted until the maintenance has finished.&lt;br /&gt;
&lt;br /&gt;
'''March 18, 2025 10:00 am EDT:''' Teach compute nodes are back.&lt;br /&gt;
&lt;br /&gt;
'''March 17, 2025 10:00 pm EDT:''' Teach compute nodes are down. We are working on it. &lt;br /&gt;
&lt;br /&gt;
'''February 27, 2025 9:00 pm EST:''' Access to HPSS via Globus has been restored.&lt;br /&gt;
&lt;br /&gt;
'''February 25, 2025 2:30 pm EST:''' Access to HPSS via Globus is currently suspended (sorry, trivial upgrade has gone wrong).&lt;br /&gt;
&lt;br /&gt;
'''February 25, 2025 12:30 pm EST:''' Mist login node is accessible again.&lt;br /&gt;
&lt;br /&gt;
'''February 25, 2025 11:50 am EST:''' Mist login node is under maintenance and temporarily inaccessible to users.&lt;br /&gt;
&lt;br /&gt;
'''February 7, 2025 2:45 pm EST:''' Systems are back online.&lt;br /&gt;
&lt;br /&gt;
'''Fri Feb  7 01:04:33 EST 2025:''' There has been a problem with the water chiller. Automatic thermal shutdown of the compute nodes&lt;br /&gt;
&lt;br /&gt;
'''January 31, 2025 11:45 am EST:''' Power is back.&lt;br /&gt;
&lt;br /&gt;
'''January 31, 2025 6:00 am EST:''' Power outage in the data center. Many compute jobs will have stopped. Until power gets restored, parts of the systems are  running on the generator. No ETA on full power restoration.&lt;br /&gt;
 &lt;br /&gt;
'''January 28, 2025 9:30 pm EST:''' The CCEnv stack has been restored.&lt;br /&gt;
&lt;br /&gt;
'''January 28, 2025 5:00 pm EST:''' The CCEnv stack from cvmfs has issues and may not work reliable.&lt;br /&gt;
&lt;br /&gt;
'''January 23, 2025 9:00 am - 1:00 pm EST:''' Balam, Rouge and Neptune compute nodes will be shut down from 9 AM to 1 PM EST for additional electrical work.&lt;br /&gt;
&lt;br /&gt;
'''January 22, 2025 12:55 pm EST:''' Compute nodes are back online&lt;br /&gt;
&lt;br /&gt;
'''January 22, 2025 8:00 am - 5:00 pm EST:''' Preparations for the new system Trillium will require a shutdown of the compute nodes of all SciNet systems (Niagara, Mist, Rouge, Teach, as well as hosted equipment) from 8 AM to 5 PM EST. The login nodes, file systems and the HPSS system will remain available. The scheduler will hold jobs that are submitted until the maintenance has finished.&lt;br /&gt;
&lt;br /&gt;
'''January 9, 2025 11:00 am EST:''' Systems are back online&lt;br /&gt;
&lt;br /&gt;
'''January 8, 2025 10:34 pm EST:''' We had some sort of thermal event at the datacenter, and the clusters are down. We're still investigating&lt;br /&gt;
&lt;br /&gt;
'''January 8, 2025 08:00 am EST:''' Balam, Rouge and Neptune are shutdown for electrical upgrades&lt;br /&gt;
&lt;br /&gt;
'''January 6, 2025:''' As part of the installation of the new computing cluster Trillium, there will be a (permanent) reduction in computing capacity of Niagara and Mist. Only 50% of Niagara and 35% of Mist will remain active after January 6th.  The reduction will require Mist to be shutdown for a few hours on January 6th. Balam, Rouge and Neptune will be shutdown on Wednesday January 8th for the same reason.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
'''December 20, 2024 09:00 am EST:''' OpenOnDemand service will not be available on Dec 20 from 9 a.m. to 5 p.m. due to scheduled maintenance.&lt;br /&gt;
&lt;br /&gt;
'''December 16, 2024, 08:21 am EST:''' The Niagara scheduler has been restarted.&lt;br /&gt;
  &lt;br /&gt;
'''December 16, 2024, 00:04 am EST:''' The Niagara scheduler has an issue; we are investigating.&lt;br /&gt;
  &lt;br /&gt;
'''Fri Nov 8, 2024, 09:45 AM EST.''' Balam and Rouge schedulers are back online.&lt;br /&gt;
&lt;br /&gt;
'''Thu Nov 7, 2024, 10:30 PM EST.''' Most systems are up, except for the schedulers on Balam and Rouge (but even their login nodes are up), and a few 'neptune' niagara nodes.&lt;br /&gt;
&lt;br /&gt;
'''Thu Nov 7, 2024, 5:30 PM EST:''' Systems are being brought up, but not yet available for users.&lt;br /&gt;
&lt;br /&gt;
'''Downtime Announcement: On Thu Nov 7, 2024, all systems and storage located at the SciNet Datacenter (Niagara, Mist, HPSS, Rouge, Teach, JupyterHub, Balam) will be unavailable from 7 a.m. to 5 p.m. ET.&lt;br /&gt;
This outage is required to install new electrical equipment (UPS) for the upcoming systems refresh. The work is expected to be completed in one day.&lt;br /&gt;
The scheduler will hold jobs that cannot finish before the start of the shutdown. Users are encouraged to submit small and short jobs that can take advantage of this, as the scheduler may be able to fit these jobs in before the maintenance on otherwise idle nodes.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 24 15:05 EDT 2024''': Cooling pump motor has been replaced. All systems are back to normal.&lt;br /&gt;
&lt;br /&gt;
'''Tue Oct 22 16:35 EDT 2024''': The motor is scheduled for replacement on Thursday, Oct 24.&lt;br /&gt;
&lt;br /&gt;
'''Mon Oct 21 17:15 EDT 2024''': Compute nodes will remain down until we can replace the main cooling pump.  This may take several days.  Please see this page for updates.&lt;br /&gt;
&lt;br /&gt;
'''Mon Oct 21 12:15 EDT 2024''': Compute nodes have been shutdown due to a cooling system failure.&lt;br /&gt;
&lt;br /&gt;
'''Fri Oct 18 21:40 EDT 2024''': Systems are back to normal&lt;br /&gt;
&lt;br /&gt;
'''Fri Oct 18 21:15 EDT 2024''': We are experiences technical difficulties, apparently caused by a glitch in the file systems&lt;br /&gt;
&lt;br /&gt;
'''Tue Oct 1 10:45 EDT 2024''': The Jupyter Hub service will be rebooted today at around 11:00 am EDT for system upgrades. &lt;br /&gt;
&lt;br /&gt;
'''Tue Sep 3 07:00 EDT 2024''': Intermittent file system issues which may cause issues logging in.  We are in the process of resolving the issue.&lt;br /&gt;
&lt;br /&gt;
'''Sun Sep 1 00:01 - 04:00 EDT 2024''': Network maintenance may cause connection issues to the datacentre.&lt;br /&gt;
&lt;br /&gt;
'''Thu Aug 22 13:30:00 EDT 2024''': Chiller issue caused about 25% of Niagara compute nodes to go down; users should resubmit any affected jobs.&lt;br /&gt;
&lt;br /&gt;
'''Wed Aug 21 16:35:00 EDT 2024''': Maintenance finished; compute nodes are now available for user jobs.&lt;br /&gt;
&lt;br /&gt;
'''Wed Aug 21 7:00:00 EDT 2024''': Maintenance started.&lt;br /&gt;
&lt;br /&gt;
'''Sun Aug 18 19:15:00 EDT 2024''': Issues have been resolved.&lt;br /&gt;
&lt;br /&gt;
'''Sun Aug 18 14:30:00 EDT 2024''': Power issues seem to have brought compute nodes down, and compounded to the file system issues we had earlier.&lt;br /&gt;
&lt;br /&gt;
'''Sun Aug 18 10:31:53 EDT 2024''': GPFS is back online, and seems to be holding&lt;br /&gt;
&lt;br /&gt;
'''Sun Aug 18 08:44:40 EDT 2024''': Sorry, problems with GPFS file systems are reoccurring. &lt;br /&gt;
&lt;br /&gt;
'''Sun Aug 18 07:59:02 EDT 2024''': GPFS file systems are back to normal. Many jobs have died and will need to be resubmitted.&lt;br /&gt;
&lt;br /&gt;
'''Sun Aug 18 06:39:12 EDT 2024''': Support staff detected the problem and started to work on the fix&lt;br /&gt;
&lt;br /&gt;
'''Sun Aug 18 00:53:52 EDT 2024''': GPFS file systems (home, scratch, project) started to show initial stages of problems&lt;br /&gt;
&lt;br /&gt;
'''August 21, 2024''': The annual cooling tower maintenance for the SciNet data centre will take place on August 21, 2024 from 7 a.m. EDT until the end of day. This maintenance requires a shutdown of the compute nodes of all SciNet systems (Niagara, Mist, Rouge, Teach, as well as hosted equipment). The login nodes, file systems and the HPSS system will remain available.&lt;br /&gt;
&lt;br /&gt;
The scheduler will hold jobs that cannot finish before the start of the shutdown. Users are encouraged to submit small and short jobs that can take advantage of this, as the scheduler may be able to fit these jobs in before the maintenance on otherwise idle nodes.&lt;br /&gt;
&lt;br /&gt;
'''Thursday, August 1, 10:00 PM EDT''' Filesystem problems resolved.&lt;br /&gt;
&lt;br /&gt;
'''Thursday, August 1, 9:30 PM EDT''' Filesystem problems preventing logins to the systems.  Working on it.&lt;br /&gt;
&lt;br /&gt;
'''Monday, July 22, 11:50 AM EDT''' Systems are back to normal&lt;br /&gt;
&lt;br /&gt;
'''Monday, July 22, 10:50 AM EDT''' Cooling problem has been fixed. Systems are coming up&lt;br /&gt;
&lt;br /&gt;
'''Monday, July 22, 10:20 AM EDT''' Compute nodes have been shutdown due to a cooling tower failure.&lt;br /&gt;
&lt;br /&gt;
'''Friday, July 19, 9:30 AM EDT''' CCEnv modules available on all login nodes again.&lt;br /&gt;
&lt;br /&gt;
'''Friday, July 19, 5:00 AM EDT''' Some login nodes do not have the CCEnv modules available.  We are working on a fix.&lt;br /&gt;
&lt;br /&gt;
'''Monday, Jun 3, 12:55 PM EDT''' All systems are recovered now.&lt;br /&gt;
&lt;br /&gt;
'''Monday, Jun 3, 10:50 AM EDT''' The file system issues affect all nodes, so all systems are inaccessible to users at the moment. No time estimate yet for when the systems may be back.&lt;br /&gt;
&lt;br /&gt;
'''Monday, Jun 3, 7:58 AM EDT''' Login issues for Niagara and Mist. There are file system issues as well. Investigating.&lt;br /&gt;
&lt;br /&gt;
'''Sunday, Jun 2, 12:00 PM EDT''' CCEnv modules missing, investigating.&lt;br /&gt;
&lt;br /&gt;
'''Wednesday May 29, 5:50 PM EDT''' Niagara compute nodes are up.  &lt;br /&gt;
&lt;br /&gt;
'''Wednesday May 29, 4:40 PM EDT''' Niagara compute nodes are coming up.  &lt;br /&gt;
&lt;br /&gt;
'''Wednesday May 29, 4 PM EDT''' Niagara login nodes and jupyterhub are up; file system is now accessible.  &lt;br /&gt;
&lt;br /&gt;
'''Wednesday May 29, 2 PM EDT''' Electricians are checking and testing all junction boxes and connectors under the raised floor for safety.  Some systems are expected to be back up later today (storage, login nodes), and compute systems will be powered up as soon as it is deemed safe.&lt;br /&gt;
&lt;br /&gt;
'''Tuesday May 28, 3 PM EDT''' Cleaning crews are at the datacentre, to pump the water and install dryers.  Once the floors are dry, we need to inspect all electrical boxes to ensure safety.  We do not expect to have a fully functional datacentre before Thursday, although we hope to be able to turn on the storage and login nodes sometime tomorrow, if circumstances permit.  Apologies, and thank you for your patience.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
'''Tuesday May 28, 7 AM EDT''' A water mains break outside our datacentre has caused extensive flooding, and all systems have been shut down preventatively. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
'''Friday May 17, 10 PM EDT - Saturday May 18, 2 AM EDT:''' The external network will be unavailable for maintenance. Running and queued jobs on the systems will not be affected.&lt;br /&gt;
&lt;br /&gt;
'''Tuesday May 14, 6:45 PM EDT:''' All systems are recovered now.&lt;br /&gt;
&lt;br /&gt;
'''Tuesday May 14, 5 PM EDT:''' Power loss at the datacentre resulted in loss of cooling.  Systems are being restored.&lt;br /&gt;
&lt;br /&gt;
'''Friday May 3, 10 PM EDT - Saturday May 4, 2 AM EDT:''' The external network will be unavailable for maintenance. Running and queued jobs on the systems will not be affected.&lt;br /&gt;
&lt;br /&gt;
'''Tuesday April 17, 2024: 11:00 ''' The restart of the Niagara login nodes has been completed successfully.&lt;br /&gt;
&lt;br /&gt;
'''Tuesday April 17, 2024: 09:40 ''' Niagara login nodes will be rebooted &lt;br /&gt;
&lt;br /&gt;
'''Tuesday April 16, 2024: 12:45 ''' mist-login01  recovered now&lt;br /&gt;
&lt;br /&gt;
'''Tuesday April 16, 2024: 11:45 ''' mist-login01  will be unavailable due to maintenance from 12:15 to 12:45. Following the completion of maintenance, login access should be restored &lt;br /&gt;
&lt;br /&gt;
'''Monday April 15, 2024: 13:02 ''' Balam-login01 will be unavailable due to maintenance from 13:00 to 13:30. Following the completion of maintenance, login access should be restored and available once more. &lt;br /&gt;
&lt;br /&gt;
'''Monday March 18, 2024: 14:45 ''' File system issue resolved.  Users are advised to check if their running jobs were affected, and if so, to resubmit.&lt;br /&gt;
&lt;br /&gt;
'''Monday March 18, 2024: 13:02 ''' File system issues.  This affects the ability to log in. We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''Monday March 11, 2024: 14:05 ''' All systems are recovered now&lt;br /&gt;
&lt;br /&gt;
'''Monday March 11, 2024:''' There will be an shutdown of the file system at SciNet for an emergency repair. As a consequence, the login nodes and compute nodes of all SciNet clusters using the file system (Niagara, Mist, Balam, Rouge, and Teach) will be down from 11 am EST until later in the afternoon. &lt;br /&gt;
&lt;br /&gt;
'''February 28, 2024, 16:30 PM EDT:''' All systems are recovered now.&lt;br /&gt;
&lt;br /&gt;
'''February 28, 2024, 1:00 PM EDT:''' A loop pump fault caused many compute nodes overheat. If you jobs failed around this time, please resubmit. Once the root cause has been addressed, the cluster will be brought up completely. Please report issues to support@scinet.utoronto.ca.&lt;br /&gt;
&lt;br /&gt;
'''February 22, 2024, 5:45 PM EDT:''' Maintenance finished and system restored. Please report issues to support@scinet.utoronto.ca.&lt;br /&gt;
&lt;br /&gt;
'''February 21, 2024, 7:00 AM EDT:''' Maintenance starting.  Niagara login nodes and the file system are kept up as much as possible, but will be rebooted at some point.&lt;br /&gt;
&lt;br /&gt;
'''February 20, 2024, 3:45 PM EDT:''' Cooling tower has been restored, all systems are in production. &lt;br /&gt;
&lt;br /&gt;
'''February 20, 2024, 1:30 AM EDT:''' Cooling tower malfunction, all compute nodes are shutdown, the root cause will be addressed earliest in the morning.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;span style=&amp;quot;color:red&amp;quot;&amp;gt;&amp;lt;b&amp;gt; February 21 and 22, 2024: SciNet Data Centre Maintenance:&amp;lt;/b&amp;gt;&amp;lt;/span&amp;gt;&amp;lt;br/&amp;gt;&lt;br /&gt;
This annual winter maintenance involves a full data centre shutdown&lt;br /&gt;
starting at 7:00 am EST on Wednesday, February 21st.  None of the&lt;br /&gt;
SciNet systems (Niagara, Mist, Rouge, Teach, the file systems, as&lt;br /&gt;
well as hosted equipment) will be accessible.  All systems should be&lt;br /&gt;
fully available again in the last afternoon of the 22nd.&lt;br /&gt;
&lt;br /&gt;
The scheduler will hold jobs that cannot finish before the start of&lt;br /&gt;
the shutdown. Users are encouraged to submit small and short jobs&lt;br /&gt;
that can take advantage of this, as the scheduler may be able to fit&lt;br /&gt;
these jobs in before the maintenance on otherwise idle nodes.&lt;br /&gt;
&lt;br /&gt;
'''Mon January 29, 08:20    (EST):''' Access to Niagara login nodes restored (it was an internal routing issue).&lt;br /&gt;
&lt;br /&gt;
'''Mon January 29, 07:35    (EST):''' No access to Niagara login nodes.  We are investigating.  Use the Mist login to get access to SciNet systems.&lt;br /&gt;
&lt;br /&gt;
'''Wed January 24, 15:20    (EST):''' maintenance on rouge-login01 &lt;br /&gt;
&lt;br /&gt;
'''Wed January 24, 14:55    (EST):''' Rebooting rouge-login01 &lt;br /&gt;
&lt;br /&gt;
'''Tue January 23, 10:25 am (EST):''' Mist-login01 maintenance done &lt;br /&gt;
&lt;br /&gt;
'''Tue January 23, 10:10 am (EST):''' Rebooting Mist-login01 to deploy new image&lt;br /&gt;
&lt;br /&gt;
'''Tue January 22, 21:00 am (EST):''' HPSS performance for hsi &amp;amp; htar clients is back to normal.&lt;br /&gt;
&lt;br /&gt;
'''Tue January 20, 11:50 am (EST):''' HPSS hsi/htar/VFS jobs will remain on PD state on the queue over the weekend, so we may work on archive02/vfs02 on Monday, and try to improve transfer performance. In the meantime you may use Globus (computecanada#hpss) if your workflow is suitable. &lt;br /&gt;
&lt;br /&gt;
'''Tue January 14, 13:20 am (EST):''' The ongoing HPSS jobs from Friday finished earlier, so we restarted HPSS sooner and released the PD jobs on the queue. &lt;br /&gt;
&lt;br /&gt;
'''Tue January 12, 10:40 am (EST):''' We have applied some tweaks to the HPSS configuration to improve performance, but they won't take effect until we restart the services, which scheduled for Monday morning. If over the weekend we notice that there are no HPSS jobs running on the queue we may restart HPSS sooner. &lt;br /&gt;
&lt;br /&gt;
'''Tue January 09, 9:10 am (EST):''' Remaining cvmfs issues cleared.&lt;br /&gt;
&lt;br /&gt;
'''Tue January 09, 8:00 am (EST):''' We're investigating remaining issues with cvmfs access on login nodes.&lt;br /&gt;
&lt;br /&gt;
'''Mon January 08, 21:50 pm (EST):''' File systems are back to normal. Please resubmit your jobs.  &lt;br /&gt;
&lt;br /&gt;
'''Mon January 08, 9:10 pm (EST):''' We had a severe deadlock, and some disk volumes went down. The file systems are being recovered now. It could take another hour.&lt;br /&gt;
&lt;br /&gt;
'''Mon January 08, 7:20 pm (EST):''' We seem to have a problem with the file system, and are investigating.&lt;br /&gt;
&lt;br /&gt;
'''Tue December 19, 2:45 pm (EST):''' Compute nodes are available again.  &lt;br /&gt;
&lt;br /&gt;
'''Tue December 19, 12:09 pm (EST):''' Maintenance was postponed by one hour. &lt;br /&gt;
&lt;br /&gt;
'''Tue December 19, 12 noon - 1 pm (EST):''' There will be a shutdown of the compute nodes of the Niagara, Mist and Rouge cluster to allow for an emergency repair to the cooling tower.  Login nodes will remain available but no jobs will run during that time.  Updates will be posted on here.&lt;br /&gt;
&lt;br /&gt;
'''Mon Dec  11 11:17:00 EST 2023:''' File systems recovered; Niagara and Mist are operational again.&lt;br /&gt;
&lt;br /&gt;
'''Mon Dec  11 7:51:00 EST 2023:''' Niagara's login nodes are being overwhelmed.  We are investigating. Likely file-system related.&lt;br /&gt;
&lt;br /&gt;
'''Thu Dec  6 10:01:24 EST 2023:''' Niagara's scheduler rebooting for security patches.&lt;br /&gt;
&lt;br /&gt;
'''Wed Dec  6 13:06:46 EST 2023:''' Endpoint computecanada#niagara transition from Globus GCSv4 to GCSv5 is completed. computecanada#niagara-GCSv4 has been deactivated&lt;br /&gt;
&lt;br /&gt;
'''Mon Dec  4 16:35:07 EST 2023:''' Endpoint computecanada#niagara has now been upgraded to Globus GCSv5. The old endpoint is still available as computecanada#niagara-GCSv4 on nia-datamover2, only until Wednesday, at which time we'll disable it as well.&lt;br /&gt;
&lt;br /&gt;
'''Mon Dec  4 11:54:49 EST 2023:''' The nia-datamover1 node will the offline this Monday afternoon for the Globus GCSv5 upgrade. Endpoint computecanada#niagara-GCSv4 will still be available via nia-datamover2&lt;br /&gt;
&lt;br /&gt;
'''Tue Nov 28 16:29:14 EST 2023:''' The computecanada#hpss Globus endpoint is now running GCSv5. We'll find a window of opportunity next week to upgrade computecanada#niagara to GCSv5 as well.&lt;br /&gt;
&lt;br /&gt;
'''Tue Nov 28 14:20:30 EST 2023:''' The computecanada#hpss Globus endpoint will be offline for the next few hours for the GCSv5 upgrade.&lt;br /&gt;
&lt;br /&gt;
'''Fri Nov 10, 2023, 18:00 PM EDT:''' The HPSS upgrade is finished. We didn't have time to update Globus to GCSv5, so we'll find a window of opportunity to do this next week. &lt;br /&gt;
&lt;br /&gt;
Please be advised that starting this &amp;lt;B&amp;gt;Friday morning, Nov/10, we'll be upgrading the HPSS system from version 8.3 to 9.3 and the HPSS Globus server from GCSv4 to GCSv5.&amp;lt;/B&amp;gt; Everything going well we expect to be back online by the end of the day.  &lt;br /&gt;
&lt;br /&gt;
'''Fri Nov 3, 2023, 12:20 PM EDT:''' The &amp;quot;Niagara at Scale&amp;quot; event has finished. Niagara is available again for all users.&lt;br /&gt;
&lt;br /&gt;
'''Tue Oct 31, 2023, 12 PM EDT:''' The &amp;quot;Niagara at Scale&amp;quot; event has started.&lt;br /&gt;
&lt;br /&gt;
'''Tue Oct 31, 2023, 12:PM EDT - Fri Nov 3, 2023, 12:00 PM EDT:''' Three-day reservation for the &amp;quot;Niagara at Scale&amp;quot; event. Only &amp;quot;Niagara at Scale&amp;quot; projects will run on the compute nodes. Users are encouraged to submit small and short jobs that could run before this event.  Throughout the event, users can still login, access their data, and submit jobs, but these jobs will not run until after the event. Note that the debugjob queue will remain available to everyone as well.&lt;br /&gt;
&lt;br /&gt;
''' Thu Oct 27 11:16 AM EDT:''' SSH keys are gradually being restored, estimated to complete by 1:15 PM.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 27, 2023, 8:00 EDT:''' SSH key login authentication with CCDB keys is currently not working, on many Alliance systems.  It appears this started last night. Issue is being investigated.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 26, 2023, 12:35 EDT:''' Mist login node is accessible again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 26, 2023, 12:05 EDT:''' Mist login node is under maintenance and temporarily inaccessible to users.&lt;br /&gt;
&lt;br /&gt;
'''Wed Oct 25 7:54 PM EDT:''' slurm-*.out now outputs job info for last array job.&lt;br /&gt;
&lt;br /&gt;
'''Tue Oct 24 12:00 AM EDT:''' network appears to be up&lt;br /&gt;
&lt;br /&gt;
'''Tue Oct 24 11:32 AM EDT:''' campus network issues&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 05, 2023, 12:05 PM EDT:''' Niagara scheduler is back online.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 05, 2023, 11:50 AM EDT:''' Niagara scheduler is temporarily under maintenance for security updates. &lt;br /&gt;
&lt;br /&gt;
''' Thu Sep 28, 2023 11:00 am''': Niagara scheduler is back online.&lt;br /&gt;
&lt;br /&gt;
''' Thu Sep 28, 2023 10:50 am''': Niagara scheduler is temporarily under maintenance for security updates.&lt;br /&gt;
&lt;br /&gt;
''' Wed Sep 27, 2023 11:35 am''': Mist login node is accessible again.&lt;br /&gt;
&lt;br /&gt;
''' Wed Sep 27, 2023 11:00 am''': Mist login node is under maintenance and temporarily inaccessible to users.&lt;br /&gt;
&lt;br /&gt;
''' Wed Sep 6, 2023 11:30 am''': Mist login node is accessible again.&lt;br /&gt;
&lt;br /&gt;
''' Wed Sep 6, 2023 11:00 am''': Mist login node is under maintenance and temporarily inaccessible to users.&lt;br /&gt;
&lt;br /&gt;
''' Fri Aug 25, 2023 0:19 am''': A power glitch brought some compute nodes down; users should resubmit any affected jobs. The Jupyterhub had to be restarted for the same reason.&lt;br /&gt;
&lt;br /&gt;
''' Mon Aug 14, 2023 12:10 pm''': Network problems with Teach cluster are now resolved and it is again available for users.&lt;br /&gt;
&lt;br /&gt;
''' Mon Aug 14, 2023 11:40 am''': Network problems with Teach cluster. We are investigating.&lt;br /&gt;
&lt;br /&gt;
''' Thu Aug 3, 2023 11:10 am''': Mist login node is accessible again.&lt;br /&gt;
&lt;br /&gt;
''' Thu Aug 3, 2023 10:40 am''': Mist login node is under maintenance and temporarily inaccessible to users.&lt;br /&gt;
&lt;br /&gt;
''' Tue Aug 1, 2023 2:43 pm''': To recover from the power glitch, all servers on the SciNet jupyterhub have been stopped. Please restart you server if you need to.&lt;br /&gt;
&lt;br /&gt;
''' Tue Aug 1, 2023 11:46 am''': There was a power glitch at 11:46 Aug 1, 2023, causing a significant number of job losses. Please resubmit your jobs.&lt;br /&gt;
&lt;br /&gt;
'''Summer Maintenance Shutdown Finished''' -- Slurm upgraded to version 23.02.3.&lt;br /&gt;
Change to be aware: SLURM_NTASKS is only set if --ntasks option is set.&lt;br /&gt;
Details at: https://bugs.schedmd.com/show_bug.cgi?id=17108&lt;br /&gt;
&lt;br /&gt;
'''July 17 and 18, 2023''':  Announcement: Summer Maintenance Shutdown&amp;lt;/span&amp;gt;&amp;lt;br&amp;gt; &lt;br /&gt;
&lt;br /&gt;
'''July 17th, 2023''' This maintenance involves a full data centre shutdown will start at 7:00 a.m. ET on Monday July 17th, 2023. None of the SciNet systems (Niagara, Mist, Rouge, Teach, the file systems, as well as hosted equipment) will be accessible.&lt;br /&gt;
&lt;br /&gt;
'''July 18th, 2023''' The shutdown will last until Tuesday July 18th, 2023. Systems are expected to be fully available in the evening of that day.&lt;br /&gt;
&lt;br /&gt;
The scheduler will hold jobs that cannot finish before the start of the shutdown. Users are encouraged to submit small and short jobs that can take advantage of this, as the scheduler may be able to fit these jobs in before the maintenance on otherwise idle nodes.&lt;br /&gt;
&lt;br /&gt;
'''Wed Jun 21 16:03:45 EDT 2023:''' Niagara's scheduler maintenance is finished.&lt;br /&gt;
&lt;br /&gt;
'''Wed Jun 21 15:42:00 EDT 2023:''' Niagara's scheduler is rebooting in 10 minutes for a short maintenance down time.&lt;br /&gt;
&lt;br /&gt;
'''Wed Jun 21, 2023, 11:25 AM EDT:''' Maintenance is finished and Teach cluster is accessible again.&lt;br /&gt;
&lt;br /&gt;
'''Tue Jun 20, 2023, 9:55 AM EDT:''' Teach cluster is powered off for maintenance.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;span style='color:red'&amp;gt;'''Tue June 20, 2023:'''  Announcement:&amp;lt;/span&amp;gt;&amp;lt;br&amp;gt; The Teach cluster at SciNet will undergo a maintenance shutdown starting on Tuesday June 20, 2023.  It will likely take a few days before it will be available again.  Check here for updates.&lt;br /&gt;
&lt;br /&gt;
'''Mon Jun 5, 2023, 2:35 PM EDT:''' All systems are operational again.&lt;br /&gt;
&lt;br /&gt;
'''Mon Jun 5, 2023, 11:55 AM EDT:''' There were issues with the cooling system.  The login nodes and file systems are now accessible again, but compute nodes are still off.&lt;br /&gt;
&lt;br /&gt;
'''Mon Jun 5, 2023, 6:55 AM EDT:''' Issues at the data center, we are investigating.&lt;br /&gt;
&lt;br /&gt;
'''Sat May 27, 2023, 21:00AM EDT:''' We have been able to mitigate the UPS issue for now, until new parts arrive sometime during the week. System will be accessible soon&lt;br /&gt;
&lt;br /&gt;
'''Sat May 27, 2023, 16:00AM EDT:''' We identified an UPS/Power related issue on the datacenter, that is adversely affecting several components, in particular all file systems. Out of an abundance of caution we are shutting down the cluster, until the UPS situation is resolved. Ongoing jobs will be canceled.&lt;br /&gt;
&lt;br /&gt;
'''Sat May 27, 2023, 11:18AM EDT:''' Filesystem issues, investigating.&lt;br /&gt;
&lt;br /&gt;
'''Wed May 24, 2023, 11:40AM EDT:''' Mist login node is accessible again.&lt;br /&gt;
&lt;br /&gt;
'''Wed May 24, 2023, 11:10 AM EDT:''' Mist login node is under maintenance and temporarily inaccessible to users.&lt;br /&gt;
&lt;br /&gt;
'''Mon May 15, 2023, 10:08 AM EDT''' rebooting Mist-login node again &lt;br /&gt;
&lt;br /&gt;
'''Mon May 15, 2023, 09:15 AM EDT''' rebooting Mist-login node&lt;br /&gt;
&lt;br /&gt;
'''Mon May 01, 2023, 04:00 PM EDT''' done rebooting nia-login nodes&lt;br /&gt;
&lt;br /&gt;
'''Mon May 01, 2023, 12:00 PM EDT''' rebooting all nia-login nodes one at a time &lt;br /&gt;
&lt;br /&gt;
'''Mon May 01, 2023, 11:00 AM EDT''' nia-login07 is going to be rebooted.&lt;br /&gt;
&lt;br /&gt;
'''Thu Apr 20, 2023, 12:05 PM EDT:''' Mist login node is accessible again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Apr 20, 2023, 11:30 AM EDT:''' Mist login node is under maintenance and temporarily inaccessible to users.&lt;br /&gt;
&lt;br /&gt;
'''Thu Apr 20, 2023, 8:27 AM EDT:''' Intermittent file system issues. We are investigating.  For now (10:45 AM), the file systems appear operational.&lt;br /&gt;
&lt;br /&gt;
'''Fri 14 Apr 2023 10:25 AM EDT:''' Switch problem resolved.&lt;br /&gt;
&lt;br /&gt;
'''Fri 14 Apr 2023 10:10 AM EDT:''' A switch problem is affecting access to certain equipment at the SciNet data center, including the Teach cluster.  Niagara and Mist are accessible.&lt;br /&gt;
&lt;br /&gt;
'''Fri 14 Apr 2023 09:55 AM EDT:''' SciNet Jupyter Hub maintenance is finished and it is again available for users.&lt;br /&gt;
&lt;br /&gt;
'''Fri 14 Apr 2023:''' SciNet Jupyter Hub will be restarted for system updates this morning.  Keep in mind to save your notebooks!&lt;br /&gt;
&lt;br /&gt;
'''Thu 06 Apr 2023 03:40 PM EDT:''' Rouge cluster is accessible again.&lt;br /&gt;
&lt;br /&gt;
'''Thu 06 Apr 2023 01:00 PM EDT:''' Rouge cluster is temporarily inaccessible to users due to the electrical work.&lt;br /&gt;
&lt;br /&gt;
'''Sun 02 Apr 2023 03:37 AM EDT:''' IO/read errors on the file system seem to have been fixed. Please resubmit your jobs, and report any further problems to support. Burst Buffer will remain offline for now.&lt;br /&gt;
&lt;br /&gt;
'''Sun 02 Apr 2023 00:18 AM EDT:''' File System is back up, but there seems to be some IO/read errors. All running jobs have been killed. Please hold off on submitting jobs until further notice.&lt;br /&gt;
&lt;br /&gt;
'''Sat 01 Apr 2023 10:17 PM EDT:''' We are having issues with the File System. Currently investigating the cause.&lt;br /&gt;
&lt;br /&gt;
'''Fri 31 Mar 2023 11:00 PM EDT:''' Burst Buffer may be the culprit. We are investigating but may have to take Burst Buffer offline. &lt;br /&gt;
&lt;br /&gt;
'''Fri 31 Mar 2023 01:30 PM EDT:''' File system issues causing trouble for some jobs on Niagara and Mist&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Tue 28 Mar 2023 11:05 AM EDT: Mist login node is accessible again.&lt;br /&gt;
&lt;br /&gt;
Tue 28 Mar 2023 10:35 AM EDT: Mist login node is under maintenance and temporarily inaccessible to users.&lt;br /&gt;
&lt;br /&gt;
Fri 17 Mar 2023 14:50 PM EDT: All systems online.&lt;br /&gt;
&lt;br /&gt;
Fri 17 Mar 2023 11:00 AM EDT: Problem identified and repaired. Starting to bring up systems, but not available to users yet.&lt;br /&gt;
&lt;br /&gt;
Fri 17 Mar 2023 09:15:39 EDT: Staff on site and ticket opened with cooling contractor, cause of failure unclear &lt;br /&gt;
&lt;br /&gt;
Fri 17 Mar 2023 01:47:43 EDT: Cooling system malfunction, datacentre is shut down. &lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Tue Feb 28, 16:40 EST:&amp;lt;/b&amp;gt; All systems are back online.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Tue Feb 28, 15:30 EST:&amp;lt;/b&amp;gt; Maintenance is complete. Bringing up systems.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Tue Feb 28, 7:10 AM EST:&amp;lt;/b&amp;gt; Maintenance shutdown resuming.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Mon Feb 27, 3:55 PM EST:&amp;lt;/b&amp;gt; Maintenance paused as parts were delayed. The maintenance will resume tomorrow (Tue Feb 28) at 7AM EST for about 5 hours.  In the meantime, the login nodes of the systems will be brought online.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Mon Feb 27, 7:20 AM EST:&amp;lt;/b&amp;gt; Maintenance shutdown started.&lt;br /&gt;
 &lt;br /&gt;
&amp;lt;span style=&amp;quot;color:red&amp;quot;&amp;gt;&amp;lt;b&amp;gt; February 27 and 28, 2023: SciNet Data Centre Maintenance:&amp;lt;/b&amp;gt;&amp;lt;/span&amp;gt;&amp;lt;br/&amp;gt;&lt;br /&gt;
This annual winter maintenance involves a full data centre shutdown&lt;br /&gt;
starting at 7:00 a.m. EST on Monday, February 27. None of the SciNet&lt;br /&gt;
systems (Niagara, Mist, Rouge, Teach, the file systems, as well as&lt;br /&gt;
hosted equipment) will be accessible.&lt;br /&gt;
&lt;br /&gt;
On the second day of the maintenance, Niagara, Mist, and their file&lt;br /&gt;
systems are expected to become partially available for users.  All&lt;br /&gt;
systems should be fully available in the evening of the 28th.&lt;br /&gt;
&lt;br /&gt;
The scheduler will hold jobs that cannot finish before the start of&lt;br /&gt;
the shutdown. Users are encouraged to submit small and short jobs&lt;br /&gt;
that can take advantage of this, as the scheduler may be able to fit&lt;br /&gt;
these jobs in before the maintenance on otherwise idle nodes.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Feb 17, 2023, 11:15 PM EST:&amp;lt;/b&amp;gt; File system issues on Teach fixed and Teach is accessible again. Note that the file system of Teach is not very good at handling many remote vscode connections.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Feb 17, 2023, 11:02 PM EST:&amp;lt;/b&amp;gt; File system issues on Teach.  We are working on a fix.  &lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Sun Feb 12, 2023, 3:05 PM EST&amp;lt;/b&amp;gt; All systems are back online.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Sun Feb 12, 2023, 2:10 PM EST&amp;lt;/b&amp;gt; Powers restored, clusters are being started.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Sat Feb 11, 2023, 2:35 PM EST&amp;lt;/b&amp;gt; Powers interruption started. All compute nodes will be down, likely until Sunday &lt;br /&gt;
afternoon.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Sat Feb 11, 2023, 1:20 PM EST&amp;lt;/b&amp;gt; There is to be an emergency power repair on the adjacent street. The datacentre will be &lt;br /&gt;
switching over to generator. All compute nodes will be down.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Fri Feb 10, 2023, 10:55 AM EST&amp;lt;/b&amp;gt; All systems are back online.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Fri Feb 10, 2023, 10:00 AM EST&amp;lt;/b&amp;gt; Cooling issue resolved, cluster is being started.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Wed Jan 25, 2023, 02:15 PM EST&amp;lt;/b&amp;gt; Mist login node is accessible again.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Wed Jan 25, 2023, 10:30 AM EST&amp;lt;/b&amp;gt; Mist login node is under maintenance and temporarily inaccessible to users.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Mon Jan 23, 2023, around 7-8 AM EST&amp;lt;/b&amp;gt; Intermediate file system issuse may have killed your job. Users are advised to resubmit.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Sat Jan 21, 2023, 00:50 EST&amp;lt;/b&amp;gt; Niagara, Mist, Rouge and the filesystems are up&lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Fri Jan 20, 2023, 11:19 PM: EST&amp;lt;/b&amp;gt; Systems are coming up. We have determined that there was a general power glitch in the area of our Datacentre. The power has been fully restored&lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Fri Jan 20, 2023, 10:34 PM: EST&amp;lt;/b&amp;gt; Cooling is back. Systems are slowly coming up  &lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Fri Jan 20, 2023, 8:20 PM: EST&amp;lt;/b&amp;gt; A cooling failure at the data center, possibly due to a power glitch. We are investigating.  &lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Thu Jan 12, 2023, 9:30 AM EST&amp;lt;/b&amp;gt; File system is experiencing issues. Issues have stabilized, but jobs running around this time may have been affected.&lt;br /&gt;
&lt;br /&gt;
'''Wed Dec 21, 2022, 12:00 PM: ''' Please note that SciNet is on vacation, together with the University of Toronto. Full service will resume on Jan 2, 2023. We will endeavour to keep systems running, and answer tickets, on a best-effort basis.  Happy Holidays!!!&lt;br /&gt;
&lt;br /&gt;
'''Fri Dec 16, 2022, 2:19 PM: ''' City power glitch caused all compute nodes to reboot. Please resubmit your jobs.&lt;br /&gt;
&lt;br /&gt;
'''Mon Dec 12, 2022, 9:30 AM - 11:30:''' File system issues caused login issues and may have affected running jobs.  System back to normal now, but users may want to check any jobs they had running. &lt;br /&gt;
&lt;br /&gt;
'''Wed Dec 7, 2022, 11:40 AM EST:''' Systems are being brought back online.&lt;br /&gt;
&lt;br /&gt;
'''Wed Dec 7, 2022, 09:00 AM EST:''' Maintenance is underway.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;span style='color:red'&amp;gt;&amp;lt;b&amp;gt;Announcement:&amp;lt;/b&amp;gt;&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
On '''Wednesday December 7th, 2022''', the file systems of the SciNet's systems, Niagara, Mist, HPSS, Teach cluster, will undergo maintenance from 9:00 am EST.  During the maintenance, there will be no access to any of these systems, as it requires all file system operations to have stopped.  The maintenance should take about 1 hour, and all systems are expected to become available again later that morning.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
'''Wed Nov 30, 2022, 14:45 PM EST:''' Mist login node is accessible again.&lt;br /&gt;
&lt;br /&gt;
'''Wed Nov 30, 2022, 14:15 PM EST:''' Mist login node is under maintenance and temporarily inaccessible to users. &lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 20, 2022, 18:00 PM EDT:''' Systems are back online &lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 20, 2022, 09:40 AM EDT:''' About half of Niagara compute nodes are up. Note that only jobs that can finish by 5:00 PM will run.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 20, 2022, 07:50 AM EDT:''' Jupyter Hub is available again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 20, 2022, 07:35 AM EDT:''' Jupyter Hub is being updated and temporarily inaccessible to users.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 20, 2022, 07:30 AM EDT:''' Maintenance is underway.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;span style='color:red'&amp;gt;&amp;lt;b&amp;gt;Announcement:&amp;lt;/b&amp;gt;&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
On '''Thursday October 20th, 2022''', the SciNet datacentre (which hosts Niagara and Mist) will undergo transformer maintenance from 7:30 am EDT to 5:00 pm EDT.  At both the start and end of this maintenance window, all systems will need to be briefly shutdown and will not be accessible.  Apart from that, during this window, login nodes will be accessible and part of Niagara will be available to run jobs. The Mist and Rouge clusters will be off for the entirety of this maintenance. &lt;br /&gt;
&lt;br /&gt;
Users are encouraged to submit Niagara jobs of about 1 to 2 hours in the days before the maintenance, as these could be run within the&lt;br /&gt;
window of 8 AM and 5 PM EDT.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
'''Wed Oct 5, 2022, 12:10 PM EDT:''' A grid power glitch caused all compute nodes to reboot. Please resubmit your jobs.&lt;br /&gt;
&lt;br /&gt;
'''Mon Oct 3, 2022, 11:20 PM EDT:'''  Niagara login nodes are accessible from outside again.&lt;br /&gt;
&lt;br /&gt;
'''Mon Oct 3, 2022, 9:20 PM EDT:'''  Niagara login nodes are inaccessible from outside of the datacentre at the moment. As a work-around, ssh into mist.scinet.utoronto.ca and then ssh into e.g. nia-login01.&lt;br /&gt;
&lt;br /&gt;
'''Wed Sep 28, 2022, 1:15 PM EDT:''' The JupyterHub maintenance is finished and it is now accessible to users.&lt;br /&gt;
&lt;br /&gt;
'''Wed Sep 28, 2022, 1:00 PM EDT:''' The JupyterHub is to be rebooted for system upgrades. Running processes and notebooks will be closed. The service is expected to be back around 1:30 PM EDT.&lt;br /&gt;
 &lt;br /&gt;
'''Tue Sep 27, 2022, 11:50 AM EDT:''' Mist login node is accessible again.&lt;br /&gt;
&lt;br /&gt;
'''Tue Sep 27, 2022, 11:25 AM EDT:''' Mist login node is under maintenance and temporarily inaccessible to users.&lt;br /&gt;
&lt;br /&gt;
'''Mon Sep 26, 2022, 11:35 AM EDT:''' Rouge and Teach login nodes are accessible again.&lt;br /&gt;
&lt;br /&gt;
'''Mon Sep 26, 2022, 11:05 AM EDT:''' Rouge and Teach login nodes are under maintenance and temporarily inaccessible to users.&lt;br /&gt;
&lt;br /&gt;
'''Fri Sep 22, 2022, 0:46 AM EDT:''' The CCEnv software stack is back to normal.&lt;br /&gt;
&lt;br /&gt;
'''Thu Sep 22, 2022, 8:15 PM EDT:''' The CCEnv software stack is inaccessible due to an issue with CVMFS.&lt;br /&gt;
 &lt;br /&gt;
'''Tue Sep 20, 2022, 16:00 AM EDT:''' Rouge login node is accessible again.&lt;br /&gt;
&lt;br /&gt;
'''Tue Sep 20, 2022, 10:20 AM EDT:''' Rouge login node is under maintenance and temporarily inaccessible to users (hardware upgrade).&lt;br /&gt;
&lt;br /&gt;
'''Tue Sep 20, 2022, 9:41 AM EDT:''' Rouge login node is back up.&lt;br /&gt;
&lt;br /&gt;
'''Tue Sep 20, 2022, 8:25 AM EDT:''' Rouge login node down, we are investigating.&lt;br /&gt;
&lt;br /&gt;
'''Fri Sept 16, 2022, 9:30 AM EDT:''' Login nodes are accessible again.&lt;br /&gt;
&lt;br /&gt;
'''Fri Sept 16, 2022, 9:00 AM EDT:''' Login nodes are not accessible.  We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''Tue Sep 13, 2022, 11:00 AM EDT:''' Mist login node is available again.&lt;br /&gt;
&lt;br /&gt;
'''Tue Sep 13, 2022, 10:00 AM EDT:''' Mist login node is under maintenance and temporarily inaccessible to users.&lt;br /&gt;
&lt;br /&gt;
'''Fri Sep 2, 2022, 11:25 AM EDT:''' Rouge login node is back up.&lt;br /&gt;
&lt;br /&gt;
'''Fri Sep 2, 2022, 10:25 AM EDT:''' Issues with the Rouge login node; we are investigating.&lt;br /&gt;
&lt;br /&gt;
'''Tue Aug 23, 2022, 1:15 PM EDT:''' Jupyter Hub is available again.&lt;br /&gt;
&lt;br /&gt;
'''Tue Aug 23, 2022, 1:00 PM EDT:''' Jupyter Hub is being updated and temporarily inaccessible to users.&lt;br /&gt;
&lt;br /&gt;
'''Fri Aug 12, 2022, 6:30 PM EDT:''' File system issues are resolved.&lt;br /&gt;
&lt;br /&gt;
'''Fri Aug 12, 2022, 5:06 PM EDT:''' File system issues. We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''Thu Aug 11, 2022, 9:20 AM EDT:''' The login node issues have been resolved.&lt;br /&gt;
&lt;br /&gt;
'''Thu Aug 11, 2022, 7:50 AM EDT:''' We are having problems accessing the Niagara login nodes.  Until fixed, please login to Mist and then ssh to a Niagara login node to access Niagara (&amp;quot;ssh nia-login02&amp;quot;, for example).&lt;br /&gt;
&lt;br /&gt;
'''Fri July 15, 2022, 10:50 AM EDT:''' Jupyter Hub is available again.&lt;br /&gt;
&lt;br /&gt;
'''Fri July 15, 2022, 10:30 AM EDT:''' Jupyter Hub is being updated and temporarily inaccessible to users.&lt;br /&gt;
&lt;br /&gt;
'''Wed June 16, 2022, 3:45 PM EDT:''' File system is stable now. We're gradually opening the systems up.&lt;br /&gt;
&lt;br /&gt;
'''Wed June 16, 2022, 10:15 AM EDT:''' Emergency maintenance shutdown of filesystem. Running jobs will be affected.&lt;br /&gt;
&lt;br /&gt;
'''Wed June 15, 2022, 7:35 PM EDT:''' Maintenance shutdown finished. Most systems are available again.&lt;br /&gt;
&lt;br /&gt;
'''Wed June 15, 2022, 7:00 AM EDT:''' Maintenance shutdown of the SciNet datacentre. There will be no access to any of the SciNet systems during this time. We expect to be able to bring the systems back online in the evening of June 15th.&lt;br /&gt;
&lt;br /&gt;
'''Mon June 13, 2022, 7:00 AM EDT - Wed June 15, 2022, 7:00 AM EDT:''' Two-day reservation for the &amp;quot;Niagara at Scale&amp;quot; event. Only &amp;quot;Niagara at Scale&amp;quot; projects will run on the compute nodes (as well as SOSCIP projects, on a subset of nodes). Users are encouraged to submit small and short jobs that could run before this event.  Throughout the event, users can still login, access their data, and submit jobs, but these jobs will not run until after the subsequent maintenance (see below). Note that the debugjob queue will remain available to everyone as well.&lt;br /&gt;
&lt;br /&gt;
'''Mon May 30th, 2022, 12:42:00 EDT:''' Mist login node is available again.&lt;br /&gt;
&lt;br /&gt;
'''Mon May 30th, 2022, 10:22:00 EDT:''' Mist login node is being upgraded and temporarily inaccessible to users.&lt;br /&gt;
&lt;br /&gt;
'''Wed May 25th, 2022, 13:30:00 EDT:''' Niagara operating at 100% again.&lt;br /&gt;
&lt;br /&gt;
'''Tue May 24th, 2022, 21:30:00 EDT:''' Jupyter Hub up.  Part of Niagara can run compute jobs.&lt;br /&gt;
&lt;br /&gt;
'''Tue May 24th, 2022, 19:00:00 EDT:''' Systems are up. Users can login, BUT cannot submit jobs yet.&lt;br /&gt;
&lt;br /&gt;
'''Tue May 24th, 2022, 10:00:00 EDT:''' We are still performing system checks.&lt;br /&gt;
&lt;br /&gt;
'''Mon May 23rd, 2022, 16:44:30 EDT:''' Systems still down. Filesystems are working, but there are quite a number of drive failures - no data loss - so out of an abundance of caution we are keeping the systems down at least until tomorrow.  The long weekend has also been disruptive for service response, and we prefer to err on the safe side.&lt;br /&gt;
&lt;br /&gt;
'''Mon May 23rd, 2022, 08:12:14 EDT:''' Systems still down. Filesystems being checked to ensure no heat damage.&lt;br /&gt;
&lt;br /&gt;
'''Sun May 22nd, 2022, 10.16 am EDT:''' Electrician dispatched to replace blown fuses.&lt;br /&gt;
&lt;br /&gt;
'''Sun May 22nd, 2022, 2:54 am EDT:''' Automatic shutdown down due to power/cooling.&lt;br /&gt;
&lt;br /&gt;
'''Fri May 6th, 2022, 11:35 am EDT:''' HPSS scheduler upgrade also finished.&lt;br /&gt;
&lt;br /&gt;
'''Thu May 5th, 2022, 7:45 pm EDT:''' Upgrade of the scheduler has finished, with the exception of HPSS.&lt;br /&gt;
&lt;br /&gt;
'''Thu May 5th, 2022, 7:00 am - 3:00 pm EDT (approx):''' Starting from 7:00 am EDT, an upgrade of the scheduler of the Niagara, Mist, and Rouge clusters will be applied.  This requires the scheduler to be down for about 5-6 hours, and all compute and login nodes to be rebooted.&lt;br /&gt;
Jobs cannot be submitted during this maintenance, but jobs submitted beforehand will remain in the queue.  For most of the time, the login nodes of the clusters will be available so that users may access their files on the home, scratch, and project file systems.&lt;br /&gt;
&lt;br /&gt;
'''Monday May 2nd, 2022, 9:30 - 11:00 am EDT:''' the Niagara login nodes, the jupyter hub, and nia-datamover2 will get rebooted for updates.  In the process, any login sessions will get disconnected, and servers on the jupyterhub will stop. Jobs in the Niagara queue will not be affected.&lt;br /&gt;
&lt;br /&gt;
'''Tue Apr 26, 11:20 AM EDT:''' A Rolling update of the Mist cluster is taking a bit longer than expected, affecting logins to Mist. &lt;br /&gt;
 &lt;br /&gt;
'''Announcement:''' On Thursday April 14th, 2022, the connectivity to the SciNet datacentre will be disrupted at 11:00 AM EDT  for a few minutes, in order to deploy a new network core switch.  Any SSH connections or data transfers to SciNet systems (Niagara, Mist, etc.) may be terminated at that time.&lt;br /&gt;
&lt;br /&gt;
'''Thu March 24, 6:54 AM EST:''' HPSS is back online&lt;br /&gt;
&lt;br /&gt;
'''Thu March 24, 8:15 AM EST:''' HPSS has a hardware problem&lt;br /&gt;
&lt;br /&gt;
'''Wed March 2, 4:50 PM EST:''' The CCEnv software stack is available again on Niagara.&lt;br /&gt;
&lt;br /&gt;
'''Wed March 2, 7:50 AM EST:''' The CCEnv software stack on Niagara has issues; we are investigating.&lt;br /&gt;
 &lt;br /&gt;
'''Sat Feb 12 2022, 12:59 EST:''' Jupyterhub is back up, but may have hardware issue.&lt;br /&gt;
&lt;br /&gt;
'''Sat Feb 12 2022, 10:36 EST:''' Issue with the Jupyterhub, since last night.  We're investigating.&lt;br /&gt;
&lt;br /&gt;
'''Tue Feb 1 2022 19:20 EST:''' Maintenance finished successfully. Systems are up. &lt;br /&gt;
&lt;br /&gt;
'''Tue Feb 1 2022 13:00 EST:''' Maintenance downtime started.&lt;br /&gt;
&lt;br /&gt;
'''Mon Jan 31 2022 13:15:00 EST:''' The SciNet datacentre's cooling system needs an '''emergency repair''' as soon as possible.  During this repair, all systems hosted at SciNet (Niagara, Mist, Rouge, HPSS, and Teach) will need to be switched off and will be unavailable to users. Repairs will start '''Tuesday February 1st, at 1:00 pm EST''', and could take until the end of the next day.  Please check here for updates.&lt;br /&gt;
&lt;br /&gt;
'''Sat Jan 29 2020 16:45:38 EST:''' Fibre repaired.&lt;br /&gt;
&lt;br /&gt;
'''Sat 29 Jan 2022 11:22:27 EST:''' Fibre repair is underway.  Expect to have connectivity restored later today.&lt;br /&gt;
&lt;br /&gt;
'''Fri 28 Jan 2022 07:35:01 EST:''' The fibre optics cable that connects the SciNet datacentre was severed by uncoordinated digging at York University.  We expect repairs to happen as soon as possible.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 27 12:46 EST PM 2022:''' Network issues to and from the datacentre. We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''Sun Jan 23 11:05 EST AM 2022:''' Filesystem issues appear to have resolved.&lt;br /&gt;
&lt;br /&gt;
'''Sun Jan 23 10:30 EST AM 2022:''' Filesystem issues -- investigating.&lt;br /&gt;
&lt;br /&gt;
'''Sat Jan 8 11:42 EST AM 2022:''' The emergency maintenance is complete. Systems are up and available.&lt;br /&gt;
&lt;br /&gt;
'''Fri Jan 7 14:34 EST PM 2022:''' The SciNet shutdown is in progress. Systems are expected back on Saturday, Jan 8.&lt;br /&gt;
&lt;br /&gt;
'''&amp;lt;span style=&amp;quot;color:red&amp;quot;&amp;gt;Emergency shutdown Friday January 7, 2022&amp;lt;/span&amp;gt;''': An emergency shutdown of all SciNet to replace a crucial file system component is planned to take place on Friday January 7, 2022, starting at 8am EST, and will require at least 12 hours of downtime.  Updates will be posted during the day.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 6 08:20 EST AM 2022''' The SciNet filesystem is having issues.  We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''Fri Dec 24 13:31 EST PM 2021''' Please note the following scheduled network maintenance, which will result in loss of connectivity to the SciNet datacentre:  Start time&lt;br /&gt;
Dec 29, 00:30 EST  Estimated duration  4 hours and 30 minutes. &lt;br /&gt;
&lt;br /&gt;
'''Mon Dec 20 4:29 EST PM 2021''' Filesystem is back to normal. &lt;br /&gt;
&lt;br /&gt;
'''Mon Dec 20 2:53 EST PM 2021''' Filesystem problem - We are investigating. &lt;br /&gt;
&lt;br /&gt;
'''Wed Sep 23 12:30 EDT 2021 ''' Cooling restored.  Systems should be available later this afternoon.  &lt;br /&gt;
&lt;br /&gt;
'''Wed Sep 23 9:30 EDT 2021 ''' Technicians on site working on cooling system. &lt;br /&gt;
&lt;br /&gt;
'''Wed Sep 23 3:30 EDT 2021 ''' Cooling system issues still unresolved. &lt;br /&gt;
&lt;br /&gt;
'''Wed Sep 22 23:27:48 EDT 2021 ''' Shutdown of the datacenter due to a problem with the cooling system.&lt;br /&gt;
&lt;br /&gt;
'''Wed Sep 22 09:30 EDT 2021 ''': File system issues, resolved.&lt;br /&gt;
&lt;br /&gt;
'''Wed Sep 22 07:30 EDT 2021 ''': File system issues, investigating.&lt;br /&gt;
&lt;br /&gt;
'''Sun Sep 19 10:00 EDT 2021''': Power glitch interrupted all compute jobs; please resubmit any jobs you had running.&lt;br /&gt;
&lt;br /&gt;
'''Wed Sep 15 17:35 EDT 2021''': filesystem issues resolved&lt;br /&gt;
&lt;br /&gt;
'''Wed Sep 15 16:39 EDT 2021''': filesystem issues&lt;br /&gt;
&lt;br /&gt;
'''Mon Sep 13 13:15:07 EDT 2021''' HPSS is back online.&lt;br /&gt;
&lt;br /&gt;
'''Fri Sep 10 17:57:23 EDT 2021''' HPSS is offline due to unscheduled maintenance.&lt;br /&gt;
&lt;br /&gt;
'''Wed Aug 18 16:13:42 EDT 2021''' The HPSS upgrade is complete.&lt;br /&gt;
&lt;br /&gt;
'''HPSS Downtime August 17th and 18th, 2021 (Tuesday and Wednesday):''' We'll be upgrading the HPSS software to version 8.3, along with all the clients (htar/hsi, vfs and Globus/dsi)&lt;br /&gt;
&lt;br /&gt;
'''July 24, 2021, 6:00 PM EDT:''' There appear to be file system issues, which may affect users' ability to login.  We are investigating.&lt;br /&gt;
&lt;br /&gt;
''' July 23th, 2021, 9:00 AM EDT:''' ''' Security update: ''' Due to a severe vulnerability in the Linux kernel (CVE-2021-33909), our team is currently patching and rebooting all login nodes and compute nodes, as well as the JupyterHub.  There should be no affect on running jobs, however sessions on login and datamover nodes will be disrupted. &lt;br /&gt;
&lt;br /&gt;
''' July 20th, 2021, 7:00 PM EDT:''' ''' SLURM configuration''' - Changed the default behaviour to kill a job step if any task exits with a non-zero exit code. If your code is able to handle failures gracefully, please add srun's option --no-kill to recover the previous default behaviour.&lt;br /&gt;
&lt;br /&gt;
''' July 20th, 2021, 7:00 PM EDT:''' Maintenance finished, systems are back online.   &lt;br /&gt;
&lt;br /&gt;
'''SciNet Downtime July 20th, 2021 (Tuesday):''' There will be a maintenance shutdown of the SciNet data center on Tuesday July 20th, starting at 7 am EDT. There will be no access to any of the SciNet systems (Niagara, Mist, HPSS, Teach cluster, or the file systems) during this time.  We expect to be able to bring the systems back online in the evening of July 20th.  The status of the Niagara cluster can be checked on status.computecanada.ca. For up-to-date and more detailed information on the status of all the SciNet systems, you can always check back here.&lt;br /&gt;
&lt;br /&gt;
'''June 28th, 2021, 4:06 PM:''' Mist OS upgrade is complete.&lt;br /&gt;
&lt;br /&gt;
'''May 27, 2021:''' Datamovers addresses have changed to improve high bandwidth connectivity and cybersecurity. The new addresses are 142.1.174.227 for nia-datamover1.scinet.utoronto.ca, and 142.1.174.228 for nia-datamover2.scinet.utoronto.ca.&lt;br /&gt;
&lt;br /&gt;
If you have jobs that need to connect to a software license server using an ssh tunnel through nia-gw (which actually resolves to datamover1 or datamover2), you may need to ask the system administrators of that license server to allow incoming connections from the new addresses above.&lt;br /&gt;
'''June 29th, 2021, 2:00 PM:''' Thunderstorm-related power fluctuations are causing some Niagara compute nodes and their jobs to crash.  Please resubmit if your jobs seem to have crashed for no apparent reason.&lt;br /&gt;
&lt;br /&gt;
'''June 28th, 2021, 4:06 PM:''' Mist OS upgrade is complete.&lt;br /&gt;
&lt;br /&gt;
'''June 28th, 2021, 9:00 AM:''' Mist is under maintenance. OS upgrading from RHEL 7 to 8.&lt;br /&gt;
&lt;br /&gt;
'''June 11th, 2021, 8:30 AM:''' Maintenance complete. Systems are up.&lt;br /&gt;
&lt;br /&gt;
'''June 9th to 10th, 2021:''' The SciNet datacentre will have a scheduled maintenance shutdown.  Niagara, Mist, Rouge, HPSS, login nodes, the file systems, and hosted systems will all be offline during the shutdown starting at 7AM EDT on Wednesday June 9th. We expect the systems to be back up in the morning of Friday June 11th.  Check here for updates.&lt;br /&gt;
&lt;br /&gt;
'''May 27, 2021:''' Datamovers addresses have changed to improve high bandwidth connectivity and cybersecurity. The new addresses are 142.1.174.227 for nia-datamover1.scinet.utoronto.ca, and 142.1.174.228 for nia-datamover2.scinet.utoronto.ca.&lt;br /&gt;
&lt;br /&gt;
'''May 27th, 20:00.''' All systems are up and running &lt;br /&gt;
&lt;br /&gt;
'''May 27th, 19:30.''' Most systems are up&lt;br /&gt;
&lt;br /&gt;
'''May 27th, 19:00:''' Cooling is back. Powering up systems&lt;br /&gt;
&lt;br /&gt;
'''May 27th, 2021, 11:30am:'''  The cooling tower issue has been identified as a wiring issue and is being repaired.  We don't have an ETA on when cooling will be restored, however we are hopeful it will be by the end of the day.  &lt;br /&gt;
&lt;br /&gt;
'''May 27th, 2021, 12:30am:''' Cooling tower motor is not working properly and may need to be replaced.  Its the primary motor and the cooling system can not run without it, so at least until tomorrow all equipment at the datacenter will remain unavailable.  Updates about expected repair times will be posted when they are known.&lt;br /&gt;
&lt;br /&gt;
'''May 26th, 2021, 9:20pm:''' we are currently experiencing cooling issues at the SciNet data centre.  Updates will be posted as we determine the cause of the problem.&lt;br /&gt;
&lt;br /&gt;
'''From Tue Mar 30 at 12 noon EST to Thu Apr 1 at 12 noon EST,''' there will be a two-day reservation for the &amp;quot;Niagara at Scale&amp;quot; pilot  event.  During these 48 hours, only &amp;quot;Niagara at Scale&amp;quot; projects will run on the compute notes (as well as SOSCIP projects, on a subset of nodes).  All other users can still login, access their data, and submit jobs throughout this event, but the jobs will not run until after the event.  The debugjob queue will remain available to  everyone as well.&lt;br /&gt;
&lt;br /&gt;
The scheduler will not start batch jobs that cannot finish before the start of this event. Users can submit small and short jobs can take advantage of this, as the scheduler may be able to fit these jobs in before the event starts on the otherwise idle nodes.&lt;br /&gt;
&lt;br /&gt;
'''Tue 23 Mar 2021 12:19:07 PM EDT''' - Planned external network maintenance 12pm-1pm Tuesday, March 23rd. &lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 28 17:35:16 EST 2021:''' HPSS services are back online&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 28 12:36:21 EST 2021:''' HPSS services offline&lt;br /&gt;
&lt;br /&gt;
We need a small maintenance window as early as possible still this afternoon to perform a small change in configuration. Ongoing jobs will be allowed to finish, but we are keeping new submissions on hold on the queue.&lt;br /&gt;
&lt;br /&gt;
'''Mon Jan 25 13:16:33 EST 2021:''' HPSS services are back online&lt;br /&gt;
&lt;br /&gt;
'''Sat Jan 23 10:03:33 EST 2021:''' HPSS services offline&lt;br /&gt;
&lt;br /&gt;
We detected some type of hardware failure on our HPSS equipment overnight, so access has been disabled pending further investigation.&lt;br /&gt;
&lt;br /&gt;
'''Fri Jan 22 10:49:29 EST 2021:''' The Globus transition to oauth is finished&lt;br /&gt;
&lt;br /&gt;
Please deactivate any previous sessions to the niagara endpoint (in the last 7 days), and activate/login again. &lt;br /&gt;
&lt;br /&gt;
For more details check https://docs.scinet.utoronto.ca/index.php/Globus#computecandada.23niagara&lt;br /&gt;
&lt;br /&gt;
'''Jan 21, 2021:''' Globus access disruption on Fri, Jan/22/2021 10AM: Please be advised that we will have a maintenance window starting tomorrow at 10AM to roll out the transition of services to oauth based authentication.&lt;br /&gt;
&lt;br /&gt;
'''Jan 15, 2021:'''Globus access update on Mon, Jan/18/2021 and Tue, Jan/19/2021:&lt;br /&gt;
Please be advised we start preparations on Monday to perform update to Globus access on Tuesday. We'll be adopting oauth instead of myproxy from that point on. During this period expect sporadic disruptions of service. On Monday we'll already block access to nia-dm2, so please refrain from starting new login sessions or ssh tunnels via nia-dm2 from this weekend already.&lt;br /&gt;
&lt;br /&gt;
''' December 11,2020, 12:00 AM EST: ''' Cooling issue resolved. Systems back.&lt;br /&gt;
&lt;br /&gt;
''' December 11,2020, 6:00 PM EST: ''' Cooling issue at datacenter. All systems down.&lt;br /&gt;
&lt;br /&gt;
''' December 7, 2020, 7:25 PM EST: '''All systems back; users can log in again.&lt;br /&gt;
&lt;br /&gt;
''' December 7, 2020, 6:46 PM EST: '''User connectivity to data center not yet ready, but queued jobs on Mist and Niagara have been started.&lt;br /&gt;
 &lt;br /&gt;
''' December 7, 2020, 7:00 AM EST: '''Maintenance shutdown in effect. This is a one-day maintenance shutdown.  There will be no access to Niagara, Mist, HPSS or teach, nor to their file systems during this time.  We expect to be able to bring the systems back online this evening.&lt;br /&gt;
&lt;br /&gt;
''' December 2, 2020, 9:10 PM EST: '''Power is back, systems are coming up. Please resubmit any jobs that failed because of this incident.&lt;br /&gt;
&lt;br /&gt;
''' December 2, 2020, 6:00 PM EST: '''Power glitch at the data center, caused about half of the compute nodes to go down.  Power issue not yet resolved.&lt;br /&gt;
&lt;br /&gt;
'''&amp;lt;span style=&amp;quot;color:#dd1111&amp;quot;&amp;gt;Announcing a Maintenance Shutdown on December 7th, 2020&amp;lt;/span&amp;gt;''' &amp;lt;br/&amp;gt;There will be a one-day maintenance shutdown on December 7th 2020, starting at 7 am EST.  There will be no access to Niagara, Mist, HPSS or teach, nor to their file systems during this time.  We expect to be able to bring the systems back online in the evening of the same day.&lt;br /&gt;
&lt;br /&gt;
''' November 6, 2020, 8:00 PM EST: ''' Systems are coming back online.&lt;br /&gt;
&lt;br /&gt;
''' November 6, 2020, 9:49 AM EST: ''' Repairs on the cooling system are underway.  No ETA, but the systems will likely be back some time today.&lt;br /&gt;
&lt;br /&gt;
''' November 6, 2020, 4:27 AM EST: '''Cooling system failure, datacentre is shut down.&lt;br /&gt;
&lt;br /&gt;
''' October 9, 2020, 12:57 PM: ''' A short power glitch caused many of the Niagara compute nodes to lose power; jobs running on them would have failed. Please check your jobs and resubmit.&lt;br /&gt;
&lt;br /&gt;
''' October 8, 2020, 9:50 PM: ''' Jupyterhub service is back up.&lt;br /&gt;
&lt;br /&gt;
''' October 8, 2020, 5:40 PM: ''' Jupyterhub service is down. We are investigating.&lt;br /&gt;
&lt;br /&gt;
''' September 28, 2020, 11:00 AM EST: ''' A short power glitch caused many of the Niagara compute nodes to lose power; jobs running on them would have failed. Please check your jobs and resubmit.&lt;br /&gt;
&lt;br /&gt;
''' September 1, 2020, 2:15 PM EST: ''' A short power glitch caused about half of the Niagara compute nodes to lose power; jobs running on them would have failed. Please check your jobs and resubmit.&lt;br /&gt;
&lt;br /&gt;
''' September 1, 2020, 9:27 AM EST: ''' The Niagara cluster has moved to a new default software stack, NiaEnv/2019b.  If your job scripts used the previous default software stack before (NiaEnv/2018a), please put the command &amp;quot;module load NiaEnv/2018a&amp;quot; before other module commands in those scripts, to ensure they will continue to work, or try the new stack (recommended).&lt;br /&gt;
''' August 24, 2020, 7:37 PM EST: ''' Connectivity is back to normal&lt;br /&gt;
&lt;br /&gt;
''' August 24, 2020, 6:35 PM EST: ''' We have partial connectivity back, but are still investigating.&lt;br /&gt;
&lt;br /&gt;
''' August 24, 2020, 3:15 PM EST: ''' There are issues connecting to the data centre. We're investigating.&lt;br /&gt;
&lt;br /&gt;
''' August 21, 2020, 6:00 PM EST: ''' The pump has been repaired, cooling is restored, systems are up.  &amp;lt;br/&amp;gt;Scratch purging is postponed until the evening of Friday Aug 28th, 2020.&lt;br /&gt;
&lt;br /&gt;
'''August 19, 2020, 4:40 PM EST:''' Update: The current estimate is to have the cooling restored on Friday and we hope to have the systems available for users on Saturday August 22, 2020.&lt;br /&gt;
&lt;br /&gt;
'''August 17, 2020, 4:00 PM EST:''' Unfortunately after taking the pump apart it was determined there was a more serious failure of the main drive shaft, not just the seal. As a new one will need to be sourced or fabricated we're estimating that it will take at least a few more days to get the part and repairs done to restore cooling. Sorry for the inconvenience. &lt;br /&gt;
&lt;br /&gt;
'''August 15, 2020, 1:00 PM EST:''' Due to parts availablity to repair the failed pump and cooling system it is unlikely that systems will be able to be restored until Monday afternoon at the earliest. &lt;br /&gt;
&lt;br /&gt;
'''August 15, 2020, 00:04 AM EST:'''  A primary pump seal in the cooling infrastructure has blown and parts availability will not be able be determined until tomorrow. All systems are shut down as there is no cooling.  If parts are available, systems may be back at the earliest late tomorrow. Check here for updates.  &lt;br /&gt;
&lt;br /&gt;
'''August 14, 2020, 21:04 AM EST:''' Tomorrow's /scratch purge has been postponed.&lt;br /&gt;
&lt;br /&gt;
'''August 14, 2020, 21:00 AM EST:''' Staff at the datacenter. Looks like one of the pumps has a seal that is leaking badly.&lt;br /&gt;
&lt;br /&gt;
'''August 14, 2020, 20:37 AM EST:''' We seem to be undergoing a thermal shutdown at the datacenter.&lt;br /&gt;
&lt;br /&gt;
'''August 14, 2020, 20:20 AM EST:''' Network problems to niagara/mist. We are investigating.&lt;br /&gt;
 &lt;br /&gt;
'''August 13, 2020, 10:40 AM EST:''' Network is fixed, scheduler and other services are back.&lt;br /&gt;
&lt;br /&gt;
'''August 13, 2020, 8:20 AM EST:''' We had an IB switch failure, which is affecting a subset of nodes, including the scheduler nodes.&lt;br /&gt;
&lt;br /&gt;
'''August 10, 2020, 7:30 PM EST:''' Scheduler fully operational again.&lt;br /&gt;
&lt;br /&gt;
'''August 10, 2020, 3:00 PM EST:''' Scheduler partially functional: jobs can be submitted and are running.&lt;br /&gt;
&lt;br /&gt;
'''August 10, 2020, 2:00 PM EST:''' Scheduler is temporarily inoperational.&lt;br /&gt;
&lt;br /&gt;
'''August 7, 2020, 9:15 PM EST:''' Network is fixed, scheduler and other services are coming back.&lt;br /&gt;
&lt;br /&gt;
'''August 7, 2020, 8:20 PM EST:''' Disruption of part of the network in the data centre.  Causes issue with the scheduler, the mist login node, and possibly others. We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''July 30, 2020, 9:00 AM''' Project backup in progress but incomplete: please be aware that after we deployed the new, larger storage appliance for scratch and project two months ago, we started a full backup of project (1.5PB). This backup is taking a while to complete, and there are still a few areas which have not been backed up fully. Please be careful to not delete things from project that you still need, in particular if they are recently added material.&lt;br /&gt;
&lt;br /&gt;
'''July 27, 2020, 5:00 PM:''' Scheduler issues resolved.&lt;br /&gt;
&lt;br /&gt;
'''July 27, 2020, 3:00 PM:''' Scheduler issues. We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''July 13, 4:40 PM:''' Most systems are available again. Only Mist is still being brought up.&lt;br /&gt;
&lt;br /&gt;
'''July 13, 10:00 AM:''' '''SciNet/Niagara Downtime In Progress'''&lt;br /&gt;
&lt;br /&gt;
'''SciNet/Niagara Downtime Announcement, July 13, 2020'''&amp;lt;br/&amp;gt;&lt;br /&gt;
All resources at SciNet will undergo a maintenance shutdown on Monday July 13, 2020, starting at 10:00 am EDT, for file system and scheduler upgrades.  There will be no access to any of the SciNet systems (Niagara, Mist, HPSS, Teach cluster, or the file systems) during this time.&lt;br /&gt;
We expect to be able to bring the systems back around 3 PM (EST) on the same day.&lt;br /&gt;
&lt;br /&gt;
''' June 29, 6:21:00  PM:''' Systems are available again.  &lt;br /&gt;
&lt;br /&gt;
''' June 29, 12:30:00  PM:''' Power Outage caused thermal shutdown.&lt;br /&gt;
&lt;br /&gt;
'''June 20, 2020, 10:24 PM:''' File systems are back up.  Unfortunately, all running jobs would have died and users are asked to resubmit them.&lt;br /&gt;
&lt;br /&gt;
'''June 20, 2020, 9:48 PM:''' An issue with the file systems is causing trouble.  We are investigating the cause.&lt;br /&gt;
&lt;br /&gt;
'''June 15, 2020, 10:30 PM:''' A '''power glitch''' caused some compute nodes to be rebooted: jobs running at the time may have failed; users are asked to resubmit these jobs.&lt;br /&gt;
&lt;br /&gt;
'''June 12, 2020, 6:15 PM:''' Two '''power glitches''' during the night caused some compute nodes to be rebooted: jobs running at the time may have failed; users are asked to resubmit these jobs.&lt;br /&gt;
&lt;br /&gt;
'''June 6, 2020, 6:06 AM:''' A '''power glitch''' caused some compute nodes to be rebooted: jobs running at the time may have failed; users are asked to resubmit these jobs.&lt;br /&gt;
&lt;br /&gt;
'''May 24, 2020, 8:20 AM:''' A '''power glitch''' this morning caused all compute nodes to be rebooted: jobs running at the time may have failed; users are asked to resubmit these jobs.&lt;br /&gt;
&lt;br /&gt;
'''May 7, 2020, 6:05 PM:''' Maintenance shutdown is finished.  Most systems are back in production.&lt;br /&gt;
&lt;br /&gt;
'''May 6, 2020, 7:08 AM:''' Two-day datacentre maintenance shutdown has started.&lt;br /&gt;
&lt;br /&gt;
''' SciNet/Niagara Downtime Announcement, May 6-7, 2020'''&lt;br /&gt;
&lt;br /&gt;
All resources at SciNet will undergo a two-day maintenance shutdown on May 6th and 7th 2020, starting at 7 am EDT on Wednesday May 6th.  There will be no access to any of the SciNet systems (Niagara, Mist, HPSS, Teach cluster, or the file systems) or systems hosted at the SciNet data centre.  We expect to be able to bring the systems back online the evening of May 7th.&lt;br /&gt;
&lt;br /&gt;
'''May 4, 2020, 7:51 AM:''' A power glitch this morning caused compute nodes to be rebooted: jobs running at the time may have failed; users are asked to resubmit these jobs.&lt;br /&gt;
&lt;br /&gt;
'''May 3, 2020, 8:20 AM:''' A power glitch this morning caused all compute nodes to be rebooted: jobs running at the time may have failed; users are asked to resubmit these jobs.&lt;br /&gt;
&lt;br /&gt;
'''April 28, 2020, 7:20 AM:''' A power glitch this morning caused all compute nodes to be rebooted: jobs running at the time have failed; users are asked to resubmit these jobs.&lt;br /&gt;
 &lt;br /&gt;
'''April 20, 2020: Security Incident at Cedar; implications for Niagara users'''&lt;br /&gt;
&lt;br /&gt;
Last week, it became evident that the Cedar GP cluster had been&lt;br /&gt;
comprimised for several weeks.  The passwords of at least two&lt;br /&gt;
Compute Canada users were known to the attackers. One of these was&lt;br /&gt;
used to escalate privileges on Cedar, as explained on&lt;br /&gt;
https://status.computecanada.ca/view_incident?incident=423.&lt;br /&gt;
&lt;br /&gt;
These accounts were used to login to Niagara as well, but Niagara&lt;br /&gt;
did not have the same security loophole as Cedar (which has been&lt;br /&gt;
fixed), and no further escalation was observed on Niagara.&lt;br /&gt;
&lt;br /&gt;
Reassuring as that may sound, it is not known how the passwords of&lt;br /&gt;
the two user accounts were obtained. Given this uncertainty, the&lt;br /&gt;
SciNet team *strongly* recommends that you change your password on&lt;br /&gt;
https://ccdb.computecanada.ca/security/change_password, and remove&lt;br /&gt;
any SSH keys and regenerate new ones (see&lt;br /&gt;
https://docs.scinet.utoronto.ca/index.php/SSH_keys).&lt;br /&gt;
&lt;br /&gt;
''' Tue 30 Mar 2020 14:55:14 EDT'''  Burst Buffer available again.&lt;br /&gt;
&lt;br /&gt;
''' Fri Mar 27 15:29:00 EDT 2020:''' SciNet systems are back up. Only the Burst Buffer remains offline, its maintenance is expected to be finished early next week.&lt;br /&gt;
&lt;br /&gt;
''' Thu Mar 26 23:05:00 EDT 2020:'''  Some aspects of the maintenance took longer than expected. The systems will not be back up until some time tomorrow, Friday March 27, 2020.  &lt;br /&gt;
&lt;br /&gt;
''' Wed Mar 25 7:00:00 EDT 2020:'''  SciNet/Niagara downtime started.&lt;br /&gt;
&lt;br /&gt;
''' Mon Mar 23 18:45:10 EDT 2020:'''  File system issues were resolved.&lt;br /&gt;
&lt;br /&gt;
''' Mon Mar 23 18:01:19 EDT 2020:''' There is currently an issue with the main Niagara filesystems. This effects all systems, all jobs have been killed. The issue is being investigated. &lt;br /&gt;
&lt;br /&gt;
''' Fri Mar 20 13:15:33 EDT 2020: ''' There was a power glitch at the datacentre at 8:50 AM, which resulted in jobs getting killed.  Please resubmit failed jobs. &lt;br /&gt;
&lt;br /&gt;
''' COVID-19 Impact on SciNet Operations, March 18, 2020'''&lt;br /&gt;
&lt;br /&gt;
Although the University of Toronto is closing of some of its&lt;br /&gt;
research operations on Friday March 20 at 5 pm EDT, this does not&lt;br /&gt;
affect the SciNet systems (such as Niagara, Mist, and HPSS), which&lt;br /&gt;
will remain operational.&lt;br /&gt;
&lt;br /&gt;
''' SciNet/Niagara Downtime Announcement, March 25-26, 2020'''&lt;br /&gt;
&lt;br /&gt;
All resources at SciNet will undergo a two-day maintenance shutdown on March 25th and 26th 2020, starting at 7 am EDT on Wednesday March 25th.  There will be no access to any of the SciNet systems (Niagara, Mist, HPSS, Teach cluster, or the file systems) during this time.&lt;br /&gt;
&lt;br /&gt;
This shutdown is necessary to finish the expansion of the Niagara cluster and its storage system.&lt;br /&gt;
&lt;br /&gt;
We expect to be able to bring the systems back online the evening of March 26th.&lt;br /&gt;
&lt;br /&gt;
''' March 9, 2020, 11:24 PM:''' HPSS services are temporarily suspended for emergency maintenance.&lt;br /&gt;
&lt;br /&gt;
''' March 7, 2020, 10:15 PM:''' File system issues have been cleared.&lt;br /&gt;
&lt;br /&gt;
''' March 6, 2020, 7:30 PM:''' File system issues; we are investigating&lt;br /&gt;
&lt;br /&gt;
''' March 2, 2020, 1:30 PM:''' For the extension of Niagara, the operating system on all Niagara nodes has been upgraded&lt;br /&gt;
from CentOS 7.4 to 7.6.  This required all&lt;br /&gt;
nodes to be rebooted. Running compute jobs are allowed to finish&lt;br /&gt;
before the compute node gets rebooted. Login nodes have all been rebooted, as have the datamover nodes and the jupyterhub service.&lt;br /&gt;
&lt;br /&gt;
''' Feb 24, 2020, 1:30PM: ''' The [[Mist]] login node got rebooted.  It is back, but we are still monitoring the situation.&lt;br /&gt;
&lt;br /&gt;
''' Feb 12, 2020, 11:00AM: ''' The [[Mist]] GPU cluster now available to users.&lt;br /&gt;
&lt;br /&gt;
''' Feb 11, 2020, 2:00PM: ''' The Niagara compute nodes were accidentally rebooted, killing all running jobs.&lt;br /&gt;
&lt;br /&gt;
''' Feb 10, 2020, 19:00PM: ''' HPSS is back to normal.&lt;br /&gt;
&lt;br /&gt;
''' Jan 30, 2020, 12:01PM: ''' We are having an issue with HPSS, in which the disk-cache is full. We put a reservation on the whole system (Globus, plus archive and vfs queues), until it has had a chance to clear some space on the cache.&lt;br /&gt;
&lt;br /&gt;
''' Jan 21, 2020, 4:05PM: '''   The was a partial power outage the took down a large amount of the compute nodes.  If your job died during this period please resubmit.  &lt;br /&gt;
&lt;br /&gt;
'''Jan 13, 2020, 7:35 PM:''' Maintenance finished.&lt;br /&gt;
&lt;br /&gt;
'''Jan 13, 2020, 8:20 AM:''' The announced maintenance downtime started (see below).&lt;br /&gt;
&lt;br /&gt;
'''Jan 9 2020, 11:30 AM:''' External ssh connectivity restored, issue related to the university network.&lt;br /&gt;
&lt;br /&gt;
'''Jan 9 2020, 9:24 AM:''' We received reports of users having trouble connecting into the SciNet data centre; we're investigating.  Systems are up and running and jobs are fine.&amp;lt;p&amp;gt;&lt;br /&gt;
As a work around, in the meantime, it appears to be possible to log into graham, cedar or beluga, and then ssh to niagara.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Downtime announcement:'''&lt;br /&gt;
To prepare for the upcoming expansion of Niagara, there will be a&lt;br /&gt;
one-day maintenance shutdown on '''January 13th 2020, starting at 8 am&lt;br /&gt;
EST'''.  There will be no access to Niagara, Mist, HPSS or teach, nor&lt;br /&gt;
to their file systems during this time.&lt;br /&gt;
&lt;br /&gt;
2019&lt;br /&gt;
&lt;br /&gt;
'''December 13, 9:00 AM EST:''' Issues resolved.&lt;br /&gt;
&lt;br /&gt;
'''December 13, 8:20 AM EST:''' Overnight issue is now preventing logins to Niagara and other services. Possibly a file system issue, we are investigating.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt; '''Fri, Nov 15 2019, 11:00 PM (EST)'''  Niagara and most of the main systems are now available. &lt;br /&gt;
&amp;lt;/p&amp;gt;&amp;lt;p&amp;gt; '''Fri, Nov 15 2019, 7:50 PM (EST)'''  SOSCIP GPU cluster is up and accessible.  Work on the other systems continues.&lt;br /&gt;
&amp;lt;/p&amp;gt;&amp;lt;p&amp;gt; '''Fri, Nov 15 2019, 5:00 PM (EST)'''  Infrastructure maintenance done, upgrades still in process.&lt;br /&gt;
&amp;lt;/p&amp;gt;&amp;lt;p&amp;gt;&lt;br /&gt;
'''Fri, Nov 15 2019, 7:00 AM (EST)'''  Maintenance shutdown of the SciNet data centre has started.  Note: scratch purging has been postponed until Nov 17.&amp;lt;br/&amp;gt; &lt;br /&gt;
&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&lt;br /&gt;
'''Announcement:''' &lt;br /&gt;
The SciNet datacentre will undergo a maintenance shutdown on&lt;br /&gt;
Friday November 15th 2019, from 7 am to 11 pm (EST), with no access&lt;br /&gt;
to any of the SciNet systems (Niagara, P8, SGC, HPSS, Teach cluster,&lt;br /&gt;
or the filesystems) during that time. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
'''Sat, Nov 2 2019, 1:30 PM (update):'''  Chiller has been fixed, all systems are operational.    &lt;br /&gt;
&amp;lt;/p&amp;gt;&lt;br /&gt;
'''Fri, Nov 1 2019, 4:30 PM (update):'''  We are operating in free cooling so have brought up about 1/2 of the Niagara compute nodes to reduce the cooling load.  Access, storage, and other systems should now be available.   &lt;br /&gt;
&lt;br /&gt;
'''Fri, Nov 1 2019, 12:05 PM (update):''' A power module in the chiller has failed and needs to be replaced.   We should be able to operate in free cooling if the temperature stays cold enough, but we may not be able to run all systems. No eta yet on when users will be able to log back in. &lt;br /&gt;
&lt;br /&gt;
'''Fri, Nov 1 2019, 9:15 AM (update):''' There was a automated shutdown because of rising temperatures, causing all systems to go down. We are investigating, check here for updates.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;'''Fri, Nov 1 2019, 8:16 AM:''' Unexpected data centre issue: Check here for updates.&lt;br /&gt;
&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
''' Thu 1 Aug 2019 5:00:00 PM ''' Systems are up and operational.   &lt;br /&gt;
&lt;br /&gt;
'''Thu 1 Aug 2019 7:00:00 AM: ''' Scheduled Downtime Maintenance of the SciNet Datacenter.  All systems will be down and unavailable starting 7am until the evening. &lt;br /&gt;
&lt;br /&gt;
'''Fri 26 Jul 2019, 16:02:26 EDT:''' There was an issue with the Burst Buffer at around 3PM, and it was recently solved. BB is OK again.&lt;br /&gt;
&lt;br /&gt;
''' Sun 30 Jun 2019 ''' The '''SOSCIP BGQ''' and '''P7''' systems were decommissioned on '''June 30th, 2019'''.  The BGQdev front end node and storage are still available.  &lt;br /&gt;
&lt;br /&gt;
'''Wed 19 Jun 2018, 1:20:00 PM:''' The BGQ is back online.&lt;br /&gt;
&lt;br /&gt;
'''Wed 19 Jun 2018, 10:00:00 AM:''' The BGQ is still down, the SOSCIP GPU nodes should be back up. &lt;br /&gt;
&lt;br /&gt;
'''Wed 19 Jun 2018, 1:40:00 AM:''' There was an issue with the SOSCIP BGQ and GPU Cluster last night about 1:42am, probably a power fluctuation that took it down.  &lt;br /&gt;
&lt;br /&gt;
'''Wed 12 Jun 2019, 3:30 AM - 7:40 AM''' Intermittent system issues on Niagara's project and scratch as the file number limit was reached. We increased the number of files allowed in total on the file system. &lt;br /&gt;
&lt;br /&gt;
'''Thu 30 May 2019, 11:00:00 PM:'''&lt;br /&gt;
The maintenance downtime of SciNet's data center has finished, and systems are being brought online now.  You can check the progress here. Some systems might not be available until Friday morning.&amp;lt;br/&amp;gt;&lt;br /&gt;
Some action on the part of users will be required when they first connect again to a Niagara login nodes or datamovers.  This is due to the security upgrade of the Niagara cluster, which is now in line with currently accepted best practices.&amp;lt;br/&amp;gt;&lt;br /&gt;
The details of the required actions can be found on the [[SSH Changes in May 2019]] wiki page.&lt;br /&gt;
&lt;br /&gt;
'''Wed 29-30 May 2019''' The SciNet datacentre will undergo a two-day maintenance shutdown, starting at 7 am EDT on Wednesday May 29th.  There will be no access to any of the SciNet systems (Niagara, P7, P8, BGQ, SGC, HPSS, Teach cluster, or the file systems) during this time.&lt;br /&gt;
&lt;br /&gt;
'''SCHEDULED SHUTDOWN''': &lt;br /&gt;
&lt;br /&gt;
Please be advised that on '''Wednesday May 29th through Thursday May 30th''', the SciNet datacentre will undergo a two-day maintenance shutdown, starting at 7 am EDT on Wednesday May 29th.  There will be no access to any of the SciNet systems (Niagara, P7, P8, BGQ, SGC, HPSS, Teach cluster, or the file systems) during this time.&lt;br /&gt;
&lt;br /&gt;
This is necessary to finish the installation of an emergency power generator, to perform the annual cooling tower maintenance, and to enhance login security.&lt;br /&gt;
&lt;br /&gt;
We expect to be able to bring the systems back online the evening of May 30th.  Due to the enhanced login security, the ssh applications of users will need to update their known host list. More detailed information on this procedure will be sent shortly before the systems are back online.&lt;br /&gt;
&lt;br /&gt;
'''Fri 5 Apr 2019:''' Software updates on Niagara: The default CCEnv software stack now uses avx512 on Niagara, and there is now a NiaEnv/2019b stack (&amp;quot;epoch&amp;quot;). &lt;br /&gt;
&lt;br /&gt;
'''Thu 4 Apr 2019:''' The 2019 compute and storage allocations have taken effect on Niagara.&lt;br /&gt;
&lt;br /&gt;
'''NOTE''':  There is scheduled network maintenance for '''Friday April 26th 12am-8am''' on the Scinet datacenter external network connection.   This will not affect internal connections and running jobs however remote connections may see interruptions during this period.&lt;br /&gt;
&lt;br /&gt;
'''Wed 24 Apr 2019 14:14 EDT:''' HPSS is back on service. Library and robot arm maintenance finished.&lt;br /&gt;
&lt;br /&gt;
'''Wed 24 Apr 2019 08:35 EDT:''' HPSS out of service this morning for library and robot arm maintenance.&lt;br /&gt;
&lt;br /&gt;
'''Fri 19 Apr 2019 17:40 EDT:''' HPSS robot arm has been released and is back to normal operations.&lt;br /&gt;
&lt;br /&gt;
'''Fri 19 Apr 2019 14:00 EDT:''' problems with HPPS library robot have been detected.&lt;br /&gt;
&lt;br /&gt;
'''Wed 17 Apr 2019 15:35 EDT:''' Network connection is back.&lt;br /&gt;
&lt;br /&gt;
'''Wed 17 Apr 2019 15:12 EDT:''' Network connection down.  Investigating.&lt;br /&gt;
&lt;br /&gt;
'''Tue 9 Apr 2019 22:24:14 EDT:'''  Network connection restored.&lt;br /&gt;
&lt;br /&gt;
'''Tue 9 Apr 2019, 15:20:''' Network connection down.  Investigating.&lt;br /&gt;
&lt;br /&gt;
'''Fri 5 Apr 2019:''' Planned, short outage in connectivity to the SciNet datacentre from 7:30 am to 8:55 am EST for maintenance of the network.  This outage will not affect running or queued jobs. It may be necessary to reboot the login nodes at some point tomorrow, which could result in a short interruption of connectivity, but which will have no effect on running or queued jobs.&lt;br /&gt;
&lt;br /&gt;
'''April 4, 2019:'''  The 2019 compute and storage allocations will take effect on Niagara. Running jobs will not be affected by this change and will run their course.  Queued jobs' priorities will be updated to reflect the new fairshare values later in the day.  The queue should fully reflect the new fairshare values in about 24 hours.   &lt;br /&gt;
&lt;br /&gt;
It may be necessary to reboot the login nodes at some point tomorrow, which could result in a short interruption of connectivity, but which will have no effect on running or queued jobs.&lt;br /&gt;
&lt;br /&gt;
There will be updates to the software stack on this day as well.&lt;br /&gt;
&lt;br /&gt;
'''March 25, 3:05 PM EST:'''  Most systems back online, other services should be back shortly. &lt;br /&gt;
&lt;br /&gt;
'''March 25, 12:05 PM EST:''' Power is back at the datacentre, but it is not yet known when all systems will be back up.  Keep checking here for updates.&lt;br /&gt;
&lt;br /&gt;
'''March 25, 11:27 AM EST:''' A power outage in the datacentre occured and caused all services to go down.  Check here for updates.&lt;br /&gt;
&lt;br /&gt;
'''Thu Mar 21 10:37:28 EDT 2019:''' HPSS is back in service&lt;br /&gt;
&lt;br /&gt;
HPSS out of service on '''Tue, Mar/19 at 9AM''', for tape library expansion and relocation. It's possible the downtime will extend to Wed, Mar/20.&lt;br /&gt;
&lt;br /&gt;
'''January 21, 4:00 PM''': HPSS is back in service. Thank you for your patience.&lt;br /&gt;
&lt;br /&gt;
'''January 18, 5:00 PM''': We did practically all of the HPSS upgrades (software/hardware), however the main client node - archive02 - is presenting an issue we just couldn't resolve yet. We will try to resume work over the weekend with cool heads, or on Monday. Sorry, but this is an unforeseen delay. Jobs on the queue we'll remain there, and we'll delay the scratch purging by 1 week.&lt;br /&gt;
&lt;br /&gt;
'''January 16, 11:00 PM''': HPSS is being upgraded, as announced.&lt;br /&gt;
&lt;br /&gt;
'''January 16, 8:00 PM''': System are coming back up and should be accessible for users now.&lt;br /&gt;
&lt;br /&gt;
'''January 15, 8:00 AM''': Data centre downtime in effect.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;font color=red&amp;gt;&amp;lt;b&amp;gt;Downtime Announcement for January 15 and 16, 2019&amp;lt;/b&amp;gt;&amp;lt;/font&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
The SciNet datacentre will need to undergo a two-day maintenance shutdown in order to perform electrical work, repairs and maintenance.  The electrical work is in preparation for the upcoming installation of an emergency power generator and a larger UPS, which will result in increased resilience to power glitches and outages.  The shutdown is scheduled to start on '''Tuesday January 15, 2019, at 7 am''' and will last until '''Wednesday 16, 2019''', some time in the evening. There will be no access to any of the SciNet systems (Niagara, P7, P8, BGQ, SGC, HPSS, Teach cluster, or the filesystems) during this time.&lt;br /&gt;
Check back here for up-to-date information on the status of the systems.&lt;br /&gt;
&lt;br /&gt;
Note: this downtime was originally scheduled for Dec. 18, 2018, but has been postponed and combined with the annual maintenance downtime.&lt;br /&gt;
&lt;br /&gt;
'''December 24, 2018, 11:35 AM EST:''' Most systems are operational again. If you had compute jobs running yesterday at around 3:30PM, they likely crashed - please check them and resubmit if needed.&lt;br /&gt;
&lt;br /&gt;
'''December 24, 2018, 10:40 AM EST:''' Repairs have been made, and the file systems are starting to be mounted on the cluster. &lt;br /&gt;
&lt;br /&gt;
'''December 23, 2018, 3:38 PM EST:''' Issues with the file systems (home, scratch and project). We are investigating, it looks like a hardware issue that we are trying to work around. Note that the absence of /home means you cannot log in with ssh keys. All compute jobs crashed around 3:30 PM EST on Dec 23. Once the system is properly up again, please resubmit your jobs.  Unfortunately, at this time of year, it is not possible to give an estimate on when the system will be operational again.&lt;br /&gt;
&lt;br /&gt;
'''Tue Nov 22 14:20:00 EDT 2018''': &amp;lt;font color=green&amp;gt;HPSS back in service&amp;lt;/font&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Tue Nov 22 08:55:00 EDT 2018''': &amp;lt;font color=red&amp;gt;HPSS offline for scheduled maintenance&amp;lt;/font&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Tue Nov 20 16:30:00 EDT 2018''':  HPSS offline on Thursday 9AM for installation of new LTO8 drives in the tape library.&lt;br /&gt;
&lt;br /&gt;
'''Tue Oct  9 12:16:00 EDT 2018''':  BGQ compute nodes are up.  &lt;br /&gt;
&lt;br /&gt;
'''Sun Oct  7 20:24:26 EDT 2018''':  SGC and BGQ front end are available,  BGQ compute nodes down related to a cooling issue.  &lt;br /&gt;
&lt;br /&gt;
'''Sat Oct  6 23:16:44 EDT 2018''':  There were some problems bringing up SGC &amp;amp; BGQ, they will remain offline for now.&lt;br /&gt;
&lt;br /&gt;
'''Sat Oct  6 18:36:35 EDT 2018''':  Electrical work finished, power restored. Systems are coming online.&lt;br /&gt;
&lt;br /&gt;
'''July 18, 2018:''' login.scinet.utoronto.ca is now disabled, GPC $SCRATCH and $HOME are decommissioned.&lt;br /&gt;
&lt;br /&gt;
'''July 12, 2018:''' There was a short power interruption around 10:30 am which caused most of the systems (Niagara, SGC, BGQ) to reboot and any running jobs to fail. &lt;br /&gt;
&lt;br /&gt;
'''July 11, 2018:''' P7's moved to BGQ filesystem, P8's moved to Niagara filesystem.&lt;br /&gt;
&lt;br /&gt;
'''May 24, 2018, 9:25 PM EST:''' The data center is up, and all systems are operational again.&lt;br /&gt;
&lt;br /&gt;
'''May 24, 2018, 7:00 AM EST:''' The data centre is under annual maintenance. All systems are offline. Systems are expected to be back late afternoon today; check for updates on this page.&lt;br /&gt;
&lt;br /&gt;
'''May 18, 2018:''' Announcement: Annual scheduled maintenance downtime: Thursday May 24, starting 7:00 AM&lt;br /&gt;
&lt;br /&gt;
'''May 16, 2018:''' Cooling  restored, systems online&lt;br /&gt;
&lt;br /&gt;
'''May 16, 2018:''' Cooling issue at datacentre again, all systems down&lt;br /&gt;
&lt;br /&gt;
'''May 15, 2018:''' Cooling restored, systems coming online&lt;br /&gt;
&lt;br /&gt;
'''May 15, 2018''' Cooling issue at datacentre, all systems down&lt;br /&gt;
&lt;br /&gt;
'''May 4, 2018:''' [[HPSS]] is now operational on Niagara.&lt;br /&gt;
&lt;br /&gt;
'''May 3, 2018:''' [[Burst Buffer]] is available upon request.&lt;br /&gt;
&lt;br /&gt;
'''May 3, 2018:''' The [https://docs.computecanada.ca/wiki/Globus Globus] endpoint for Niagara is available: computecanada#niagara.&lt;br /&gt;
&lt;br /&gt;
'''May 1, 2018:''' System status moved he here.&lt;br /&gt;
&lt;br /&gt;
'''Apr 23, 2018:''' GPC-compute is decommissioned, GPC-storage available until 30 May 2018.&lt;br /&gt;
&lt;br /&gt;
'''April 10, 2018:''' Niagara commissioned.&lt;/div&gt;</summary>
		<author><name>Rzon</name></author>
	</entry>
	<entry>
		<id>https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7706</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7706"/>
		<updated>2026-05-01T15:57:56Z</updated>

		<summary type="html">&lt;p&gt;Rzon: /* System Status */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;__NOTOC__&lt;br /&gt;
{| style=&amp;quot;border-spacing:10px; width: 95%&amp;quot;&lt;br /&gt;
| style=&amp;quot;padding:1em; padding-top:.1em; border:2px solid #0645ad; background-color:#f6f6f6; border-radius:7px&amp;quot;|&lt;br /&gt;
&lt;br /&gt;
==System Status==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- Use &amp;quot;Up&amp;quot;, &amp;quot;Partial&amp;quot; or &amp;quot;Down&amp;quot;; these are templates. --&amp;gt;&lt;br /&gt;
{|style=&amp;quot;width:100%&amp;quot; &lt;br /&gt;
|{{Up3 | Trillium|https://docs.alliancecan.ca/wiki/Trillium_Quickstart}}&lt;br /&gt;
|{{Up3 | OnDemand|https://docs.alliancecan.ca/wiki/Trillium_Open_OnDemand_Quickstart}}&lt;br /&gt;
|{{Up | Globus |Globus}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Up | HPSS|HPSS}}&lt;br /&gt;
|{{up | Balam|Balam}}&lt;br /&gt;
|{{Up | S4H | S4H}}&lt;br /&gt;
|-&lt;br /&gt;
|{{up | Teach|Teach}}&lt;br /&gt;
|{{Up3 | File system|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Storage}}&lt;br /&gt;
|{{Up3 | External Network|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Logging_in}} &lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Wed Apr 30, 2026, 3:00 pm:''' Most system have been updated to mitigate known security risks, and are back in service. Note that no actual security breaches were found.&lt;br /&gt;
&lt;br /&gt;
'''Wed Apr 29, 2026, 5:25 pm:''' For security reasons, login access to all systems has been disabled, as have OnDemand Apps.  Compute jobs are still allowed to run.&lt;br /&gt;
 &lt;br /&gt;
'''Thu Apr 23, 2026, 10:00 am:''' The Trillium file system has issues and may be slow on certain nodes. We are still investigating.&lt;br /&gt;
&lt;br /&gt;
'''Tue Apr 14, 2026, 10:30 am:''' The Trillium globus endpoint is working again.&lt;br /&gt;
&lt;br /&gt;
'''Sat Apr 11, 2026, 10:00 pm:''' The Trillium globus endpoint is not operational (it times out). We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''Thu Apr 09, 2026, 10:30 am:''' tri-dm4.scinet.utoronto.ca and robot4.scinet.utoronto.ca are in maintenance. Use 1,2, or 3 instead.&lt;br /&gt;
&lt;br /&gt;
'''Wed Apr 08, 2026, 6:30 pm:''' Software from the CVMFS 'restricted' area is back.&lt;br /&gt;
&lt;br /&gt;
'''Wed Apr 08, 2026, 1:40 pm:''' Software from the CVMFS 'restricted' area is not available on many of the Trillium and Open OnDemand nodes. We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''Mon Apr 06, 2026, 10:00 pm:''' We will have to reschedule the HPSS update. This attempt didn't work as expected&lt;br /&gt;
&lt;br /&gt;
'''Mon Apr 06, 2026, 8:00 pm:''' HPSS scheduled maintenance: update of HPSS to v11.3_u4 and hsi-htar to v11.3_u1 (bug fixes)&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--  When removing system status entries, please archive them to: --&amp;gt;&lt;br /&gt;
[[Previous messages]]&lt;br /&gt;
&lt;br /&gt;
{|style=&amp;quot;border-spacing: 10px;width: 100%&amp;quot;&lt;br /&gt;
|valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== QuickStart Guides ==&lt;br /&gt;
* [https://docs.alliancecan.ca/wiki/Trillium_Quickstart Trillium Quickstart]&lt;br /&gt;
* [[HPSS | HPSS archival storage]]&lt;br /&gt;
* [[Teach|Teach cluster]]&lt;br /&gt;
* [[FAQ | FAQ (frequently asked questions)]]&lt;br /&gt;
* [[Acknowledging SciNet]]&lt;br /&gt;
| valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== Tutorials, Manuals, etc. ==&lt;br /&gt;
* [https://education.scinet.utoronto.ca SciNet education material]&lt;br /&gt;
* [https://www.youtube.com/c/SciNetHPCattheUniversityofToronto SciNet's YouTube channel]&lt;br /&gt;
* [[Modules for Mist]] &lt;br /&gt;
* [[Commercial software]]&lt;br /&gt;
* [[Burst Buffer]]&lt;br /&gt;
* [[SSH#SSH Keys|SSH keys]]&lt;br /&gt;
* [[SSH Tunneling]]&lt;br /&gt;
* [[Visualization]]&lt;br /&gt;
* [[Running Serial Jobs on Niagara]]&lt;br /&gt;
* [[Jupyter Hub]]&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Rzon</name></author>
	</entry>
	<entry>
		<id>https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7703</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7703"/>
		<updated>2026-05-01T13:32:26Z</updated>

		<summary type="html">&lt;p&gt;Rzon: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;__NOTOC__&lt;br /&gt;
{| style=&amp;quot;border-spacing:10px; width: 95%&amp;quot;&lt;br /&gt;
| style=&amp;quot;padding:1em; padding-top:.1em; border:2px solid #0645ad; background-color:#f6f6f6; border-radius:7px&amp;quot;|&lt;br /&gt;
&lt;br /&gt;
==System Status==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- Use &amp;quot;Up&amp;quot;, &amp;quot;Partial&amp;quot; or &amp;quot;Down&amp;quot;; these are templates. --&amp;gt;&lt;br /&gt;
{|style=&amp;quot;width:100%&amp;quot; &lt;br /&gt;
|{{Up3 | Trillium|https://docs.alliancecan.ca/wiki/Trillium_Quickstart}}&lt;br /&gt;
|{{Up3 | OnDemand|https://docs.alliancecan.ca/wiki/Trillium_Open_OnDemand_Quickstart}}&lt;br /&gt;
|{{Up | Globus |Globus}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Down | HPSS|HPSS}}&lt;br /&gt;
|{{up | Balam|Balam}}&lt;br /&gt;
|{{Up | S4H | S4H}}&lt;br /&gt;
|-&lt;br /&gt;
|{{up | Teach|Teach}}&lt;br /&gt;
|{{Up3 | File system|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Storage}}&lt;br /&gt;
|{{Up3 | External Network|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Logging_in}} &lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Wed Apr 30, 2026, 3:00 pm:''' Most system have been updated to mitigate known security risks, and are back in service. Note that no actual security breaches were found.&lt;br /&gt;
&lt;br /&gt;
'''Wed Apr 29, 2026, 5:25 pm:''' For security reasons, login access to all systems has been disabled, as have OnDemand Apps.  Compute jobs are still allowed to run.&lt;br /&gt;
 &lt;br /&gt;
'''Thu Apr 23, 2026, 10:00 am:''' The Trillium file system has issues and may be slow on certain nodes. We are still investigating.&lt;br /&gt;
&lt;br /&gt;
'''Tue Apr 14, 2026, 10:30 am:''' The Trillium globus endpoint is working again.&lt;br /&gt;
&lt;br /&gt;
'''Sat Apr 11, 2026, 10:00 pm:''' The Trillium globus endpoint is not operational (it times out). We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''Thu Apr 09, 2026, 10:30 am:''' tri-dm4.scinet.utoronto.ca and robot4.scinet.utoronto.ca are in maintenance. Use 1,2, or 3 instead.&lt;br /&gt;
&lt;br /&gt;
'''Wed Apr 08, 2026, 6:30 pm:''' Software from the CVMFS 'restricted' area is back.&lt;br /&gt;
&lt;br /&gt;
'''Wed Apr 08, 2026, 1:40 pm:''' Software from the CVMFS 'restricted' area is not available on many of the Trillium and Open OnDemand nodes. We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''Mon Apr 06, 2026, 10:00 pm:''' We will have to reschedule the HPSS update. This attempt didn't work as expected&lt;br /&gt;
&lt;br /&gt;
'''Mon Apr 06, 2026, 8:00 pm:''' HPSS scheduled maintenance: update of HPSS to v11.3_u4 and hsi-htar to v11.3_u1 (bug fixes)&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--  When removing system status entries, please archive them to: --&amp;gt;&lt;br /&gt;
[[Previous messages]]&lt;br /&gt;
&lt;br /&gt;
{|style=&amp;quot;border-spacing: 10px;width: 100%&amp;quot;&lt;br /&gt;
|valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== QuickStart Guides ==&lt;br /&gt;
* [https://docs.alliancecan.ca/wiki/Trillium_Quickstart Trillium Quickstart]&lt;br /&gt;
* [[HPSS | HPSS archival storage]]&lt;br /&gt;
* [[Teach|Teach cluster]]&lt;br /&gt;
* [[FAQ | FAQ (frequently asked questions)]]&lt;br /&gt;
* [[Acknowledging SciNet]]&lt;br /&gt;
| valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== Tutorials, Manuals, etc. ==&lt;br /&gt;
* [https://education.scinet.utoronto.ca SciNet education material]&lt;br /&gt;
* [https://www.youtube.com/c/SciNetHPCattheUniversityofToronto SciNet's YouTube channel]&lt;br /&gt;
* [[Modules for Mist]] &lt;br /&gt;
* [[Commercial software]]&lt;br /&gt;
* [[Burst Buffer]]&lt;br /&gt;
* [[SSH#SSH Keys|SSH keys]]&lt;br /&gt;
* [[SSH Tunneling]]&lt;br /&gt;
* [[Visualization]]&lt;br /&gt;
* [[Running Serial Jobs on Niagara]]&lt;br /&gt;
* [[Jupyter Hub]]&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Rzon</name></author>
	</entry>
	<entry>
		<id>https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7694</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7694"/>
		<updated>2026-04-30T19:08:58Z</updated>

		<summary type="html">&lt;p&gt;Rzon: /* System Status */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;__NOTOC__&lt;br /&gt;
{| style=&amp;quot;border-spacing:10px; width: 95%&amp;quot;&lt;br /&gt;
| style=&amp;quot;padding:1em; padding-top:.1em; border:2px solid #0645ad; background-color:#f6f6f6; border-radius:7px&amp;quot;|&lt;br /&gt;
&lt;br /&gt;
==System Status==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- Use &amp;quot;Up&amp;quot;, &amp;quot;Partial&amp;quot; or &amp;quot;Down&amp;quot;; these are templates. --&amp;gt;&lt;br /&gt;
{|style=&amp;quot;width:100%&amp;quot; &lt;br /&gt;
|{{Up3 | Trillium|https://docs.alliancecan.ca/wiki/Trillium_Quickstart}}&lt;br /&gt;
|{{Down3 | OnDemand|https://docs.alliancecan.ca/wiki/Trillium_Open_OnDemand_Quickstart}}&lt;br /&gt;
|{{Down | Globus |Globus}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Down | HPSS|HPSS}}&lt;br /&gt;
|{{up | Balam|Balam}}&lt;br /&gt;
|{{Partial | S4H | S4H}}&lt;br /&gt;
|-&lt;br /&gt;
|{{up | Teach|Teach}}&lt;br /&gt;
|{{Up3 | File system|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Storage}}&lt;br /&gt;
|{{Up3 | External Network|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Logging_in}} &lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Wed Apr 30, 2026, 3:00 pm:''' Most systems are back.&lt;br /&gt;
&lt;br /&gt;
'''Wed Apr 29, 2026, 5:25 pm:''' For security reasons, login access to all systems has been disabled, as have OnDemand Apps.  Compute jobs are still allowed to run.&lt;br /&gt;
 &lt;br /&gt;
'''Thu Apr 23, 2026, 10:00 am:''' The Trillium file system has issues and may be slow on certain nodes. We are still investigating.&lt;br /&gt;
&lt;br /&gt;
'''Tue Apr 14, 2026, 10:30 am:''' The Trillium globus endpoint is working again.&lt;br /&gt;
&lt;br /&gt;
'''Sat Apr 11, 2026, 10:00 pm:''' The Trillium globus endpoint is not operational (it times out). We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''Thu Apr 09, 2026, 10:30 am:''' tri-dm4.scinet.utoronto.ca and robot4.scinet.utoronto.ca are in maintenance. Use 1,2, or 3 instead.&lt;br /&gt;
&lt;br /&gt;
'''Wed Apr 08, 2026, 6:30 pm:''' Software from the CVMFS 'restricted' area is back.&lt;br /&gt;
&lt;br /&gt;
'''Wed Apr 08, 2026, 1:40 pm:''' Software from the CVMFS 'restricted' area is not available on many of the Trillium and Open OnDemand nodes. We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''Mon Apr 06, 2026, 10:00 pm:''' We will have to reschedule the HPSS update. This attempt didn't work as expected&lt;br /&gt;
&lt;br /&gt;
'''Mon Apr 06, 2026, 8:00 pm:''' HPSS scheduled maintenance: update of HPSS to v11.3_u4 and hsi-htar to v11.3_u1 (bug fixes)&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--  When removing system status entries, please archive them to: --&amp;gt;&lt;br /&gt;
[[Previous messages]]&lt;br /&gt;
&lt;br /&gt;
{|style=&amp;quot;border-spacing: 10px;width: 100%&amp;quot;&lt;br /&gt;
|valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== QuickStart Guides ==&lt;br /&gt;
* [https://docs.alliancecan.ca/wiki/Trillium_Quickstart Trillium Quickstart]&lt;br /&gt;
* [[HPSS | HPSS archival storage]]&lt;br /&gt;
* [[Teach|Teach cluster]]&lt;br /&gt;
* [[FAQ | FAQ (frequently asked questions)]]&lt;br /&gt;
* [[Acknowledging SciNet]]&lt;br /&gt;
| valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== Tutorials, Manuals, etc. ==&lt;br /&gt;
* [https://education.scinet.utoronto.ca SciNet education material]&lt;br /&gt;
* [https://www.youtube.com/c/SciNetHPCattheUniversityofToronto SciNet's YouTube channel]&lt;br /&gt;
* [[Modules for Mist]] &lt;br /&gt;
* [[Commercial software]]&lt;br /&gt;
* [[Burst Buffer]]&lt;br /&gt;
* [[SSH#SSH Keys|SSH keys]]&lt;br /&gt;
* [[SSH Tunneling]]&lt;br /&gt;
* [[Visualization]]&lt;br /&gt;
* [[Running Serial Jobs on Niagara]]&lt;br /&gt;
* [[Jupyter Hub]]&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Rzon</name></author>
	</entry>
	<entry>
		<id>https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7691</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7691"/>
		<updated>2026-04-30T18:56:15Z</updated>

		<summary type="html">&lt;p&gt;Rzon: /* System Status */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;__NOTOC__&lt;br /&gt;
{| style=&amp;quot;border-spacing:10px; width: 95%&amp;quot;&lt;br /&gt;
| style=&amp;quot;padding:1em; padding-top:.1em; border:2px solid #0645ad; background-color:#f6f6f6; border-radius:7px&amp;quot;|&lt;br /&gt;
&lt;br /&gt;
==System Status==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- Use &amp;quot;Up&amp;quot;, &amp;quot;Partial&amp;quot; or &amp;quot;Down&amp;quot;; these are templates. --&amp;gt;&lt;br /&gt;
{|style=&amp;quot;width:100%&amp;quot; &lt;br /&gt;
|{{Up3 | Trillium|https://docs.alliancecan.ca/wiki/Trillium_Quickstart}}&lt;br /&gt;
|{{Down3 | OnDemand|https://docs.alliancecan.ca/wiki/Trillium_Open_OnDemand_Quickstart}}&lt;br /&gt;
|{{Down | Globus |Globus}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Down | HPSS|HPSS}}&lt;br /&gt;
|{{up | Balam|Balam}}&lt;br /&gt;
|{{Partial | S4H | S4H}}&lt;br /&gt;
|-&lt;br /&gt;
|{{up | Teach|Teach}}&lt;br /&gt;
|{{Up3 | File system|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Storage}}&lt;br /&gt;
|{{Up3 | External Network|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Logging_in}} &lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Wed Apr 29, 2026, 5:25 pm:''' For security reasons, login access to all systems has been disabled, as have OnDemand Apps.  Compute jobs are still allowed to run.&lt;br /&gt;
 &lt;br /&gt;
'''Thu Apr 23, 2026, 10:00 am:''' The Trillium file system has issues and may be slow on certain nodes. We are still investigating.&lt;br /&gt;
&lt;br /&gt;
'''Tue Apr 14, 2026, 10:30 am:''' The Trillium globus endpoint is working again.&lt;br /&gt;
&lt;br /&gt;
'''Sat Apr 11, 2026, 10:00 pm:''' The Trillium globus endpoint is not operational (it times out). We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''Thu Apr 09, 2026, 10:30 am:''' tri-dm4.scinet.utoronto.ca and robot4.scinet.utoronto.ca are in maintenance. Use 1,2, or 3 instead.&lt;br /&gt;
&lt;br /&gt;
'''Wed Apr 08, 2026, 6:30 pm:''' Software from the CVMFS 'restricted' area is back.&lt;br /&gt;
&lt;br /&gt;
'''Wed Apr 08, 2026, 1:40 pm:''' Software from the CVMFS 'restricted' area is not available on many of the Trillium and Open OnDemand nodes. We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''Mon Apr 06, 2026, 10:00 pm:''' We will have to reschedule the HPSS update. This attempt didn't work as expected&lt;br /&gt;
&lt;br /&gt;
'''Mon Apr 06, 2026, 8:00 pm:''' HPSS scheduled maintenance: update of HPSS to v11.3_u4 and hsi-htar to v11.3_u1 (bug fixes)&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--  When removing system status entries, please archive them to: --&amp;gt;&lt;br /&gt;
[[Previous messages]]&lt;br /&gt;
&lt;br /&gt;
{|style=&amp;quot;border-spacing: 10px;width: 100%&amp;quot;&lt;br /&gt;
|valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== QuickStart Guides ==&lt;br /&gt;
* [https://docs.alliancecan.ca/wiki/Trillium_Quickstart Trillium Quickstart]&lt;br /&gt;
* [[HPSS | HPSS archival storage]]&lt;br /&gt;
* [[Teach|Teach cluster]]&lt;br /&gt;
* [[FAQ | FAQ (frequently asked questions)]]&lt;br /&gt;
* [[Acknowledging SciNet]]&lt;br /&gt;
| valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== Tutorials, Manuals, etc. ==&lt;br /&gt;
* [https://education.scinet.utoronto.ca SciNet education material]&lt;br /&gt;
* [https://www.youtube.com/c/SciNetHPCattheUniversityofToronto SciNet's YouTube channel]&lt;br /&gt;
* [[Modules for Mist]] &lt;br /&gt;
* [[Commercial software]]&lt;br /&gt;
* [[Burst Buffer]]&lt;br /&gt;
* [[SSH#SSH Keys|SSH keys]]&lt;br /&gt;
* [[SSH Tunneling]]&lt;br /&gt;
* [[Visualization]]&lt;br /&gt;
* [[Running Serial Jobs on Niagara]]&lt;br /&gt;
* [[Jupyter Hub]]&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Rzon</name></author>
	</entry>
	<entry>
		<id>https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7676</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7676"/>
		<updated>2026-04-29T21:35:26Z</updated>

		<summary type="html">&lt;p&gt;Rzon: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;__NOTOC__&lt;br /&gt;
{| style=&amp;quot;border-spacing:10px; width: 95%&amp;quot;&lt;br /&gt;
| style=&amp;quot;padding:1em; padding-top:.1em; border:2px solid #0645ad; background-color:#f6f6f6; border-radius:7px&amp;quot;|&lt;br /&gt;
&lt;br /&gt;
==System Status==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- Use &amp;quot;Up&amp;quot;, &amp;quot;Partial&amp;quot; or &amp;quot;Down&amp;quot;; these are templates. --&amp;gt;&lt;br /&gt;
{|style=&amp;quot;width:100%&amp;quot; &lt;br /&gt;
|{{Partial3 | Trillium|https://docs.alliancecan.ca/wiki/Trillium_Quickstart}}&lt;br /&gt;
|{{Partial3 | OnDemand|https://docs.alliancecan.ca/wiki/Trillium_Open_OnDemand_Quickstart}}&lt;br /&gt;
|{{Down | Globus |Globus}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Down | HPSS|HPSS}}&lt;br /&gt;
|{{Partial | Balam|Balam}}&lt;br /&gt;
|{{Partial | S4H | S4H}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Down | Teach|Teach}}&lt;br /&gt;
|{{Up3 | File system|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Storage}}&lt;br /&gt;
|{{Up3 | External Network|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Logging_in}} &lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Wed Apr 29, 2026, 5:25 pm:''' For security reasons, login access to all systems has been disabled, as have OnDemand Apps.  Compute jobs are still allowed to run.&lt;br /&gt;
 &lt;br /&gt;
'''Thu Apr 23, 2026, 10:00 am:''' The Trillium file system has issues and may be slow on certain nodes. We are still investigating.&lt;br /&gt;
&lt;br /&gt;
'''Tue Apr 14, 2026, 10:30 am:''' The Trillium globus endpoint is working again.&lt;br /&gt;
&lt;br /&gt;
'''Sat Apr 11, 2026, 10:00 pm:''' The Trillium globus endpoint is not operational (it times out). We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''Thu Apr 09, 2026, 10:30 am:''' tri-dm4.scinet.utoronto.ca and robot4.scinet.utoronto.ca are in maintenance. Use 1,2, or 3 instead.&lt;br /&gt;
&lt;br /&gt;
'''Wed Apr 08, 2026, 6:30 pm:''' Software from the CVMFS 'restricted' area is back.&lt;br /&gt;
&lt;br /&gt;
'''Wed Apr 08, 2026, 1:40 pm:''' Software from the CVMFS 'restricted' area is not available on many of the Trillium and Open OnDemand nodes. We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''Mon Apr 06, 2026, 10:00 pm:''' We will have to reschedule the HPSS update. This attempt didn't work as expected&lt;br /&gt;
&lt;br /&gt;
'''Mon Apr 06, 2026, 8:00 pm:''' HPSS scheduled maintenance: update of HPSS to v11.3_u4 and hsi-htar to v11.3_u1 (bug fixes)&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--  When removing system status entries, please archive them to: --&amp;gt;&lt;br /&gt;
[[Previous messages]]&lt;br /&gt;
&lt;br /&gt;
{|style=&amp;quot;border-spacing: 10px;width: 100%&amp;quot;&lt;br /&gt;
|valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== QuickStart Guides ==&lt;br /&gt;
* [https://docs.alliancecan.ca/wiki/Trillium_Quickstart Trillium Quickstart]&lt;br /&gt;
* [[HPSS | HPSS archival storage]]&lt;br /&gt;
* [[Teach|Teach cluster]]&lt;br /&gt;
* [[FAQ | FAQ (frequently asked questions)]]&lt;br /&gt;
* [[Acknowledging SciNet]]&lt;br /&gt;
| valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== Tutorials, Manuals, etc. ==&lt;br /&gt;
* [https://education.scinet.utoronto.ca SciNet education material]&lt;br /&gt;
* [https://www.youtube.com/c/SciNetHPCattheUniversityofToronto SciNet's YouTube channel]&lt;br /&gt;
* [[Modules for Mist]] &lt;br /&gt;
* [[Commercial software]]&lt;br /&gt;
* [[Burst Buffer]]&lt;br /&gt;
* [[SSH#SSH Keys|SSH keys]]&lt;br /&gt;
* [[SSH Tunneling]]&lt;br /&gt;
* [[Visualization]]&lt;br /&gt;
* [[Running Serial Jobs on Niagara]]&lt;br /&gt;
* [[Jupyter Hub]]&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Rzon</name></author>
	</entry>
	<entry>
		<id>https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7673</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7673"/>
		<updated>2026-04-29T21:30:10Z</updated>

		<summary type="html">&lt;p&gt;Rzon: /* System Status */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;__NOTOC__&lt;br /&gt;
{| style=&amp;quot;border-spacing:10px; width: 95%&amp;quot;&lt;br /&gt;
| style=&amp;quot;padding:1em; padding-top:.1em; border:2px solid #0645ad; background-color:#f6f6f6; border-radius:7px&amp;quot;|&lt;br /&gt;
&lt;br /&gt;
==System Status==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- Use &amp;quot;Up&amp;quot;, &amp;quot;Partial&amp;quot; or &amp;quot;Down&amp;quot;; these are templates. --&amp;gt;&lt;br /&gt;
{|style=&amp;quot;width:100%&amp;quot; &lt;br /&gt;
|{{Partial3 | Trillium|https://docs.alliancecan.ca/wiki/Trillium_Quickstart}}&lt;br /&gt;
|{{Partial3 | OnDemand|https://docs.alliancecan.ca/wiki/Trillium_Open_OnDemand_Quickstart}}&lt;br /&gt;
|{{Down | Globus |Globus}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Down | HPSS|HPSS}}&lt;br /&gt;
|{{Partial | Balam|Balam}}&lt;br /&gt;
|{{Partial | S4H | S4H}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Down | Teach|Teach}}&lt;br /&gt;
|{{Up3 | File system|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Storage}}&lt;br /&gt;
|{{Up3 | External Network|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Logging_in}} &lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Wed Apr 29, 2026, 5:25 pm:''' For security reasons, login access to all systems has been disabled.  Compute jobs are still allowed to run.&lt;br /&gt;
 &lt;br /&gt;
'''Thu Apr 23, 2026, 10:00 am:''' The Trillium file system has issues and may be slow on certain nodes. We are still investigating.&lt;br /&gt;
&lt;br /&gt;
'''Tue Apr 14, 2026, 10:30 am:''' The Trillium globus endpoint is working again.&lt;br /&gt;
&lt;br /&gt;
'''Sat Apr 11, 2026, 10:00 pm:''' The Trillium globus endpoint is not operational (it times out). We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''Thu Apr 09, 2026, 10:30 am:''' tri-dm4.scinet.utoronto.ca and robot4.scinet.utoronto.ca are in maintenance. Use 1,2, or 3 instead.&lt;br /&gt;
&lt;br /&gt;
'''Wed Apr 08, 2026, 6:30 pm:''' Software from the CVMFS 'restricted' area is back.&lt;br /&gt;
&lt;br /&gt;
'''Wed Apr 08, 2026, 1:40 pm:''' Software from the CVMFS 'restricted' area is not available on many of the Trillium and Open OnDemand nodes. We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''Mon Apr 06, 2026, 10:00 pm:''' We will have to reschedule the HPSS update. This attempt didn't work as expected&lt;br /&gt;
&lt;br /&gt;
'''Mon Apr 06, 2026, 8:00 pm:''' HPSS scheduled maintenance: update of HPSS to v11.3_u4 and hsi-htar to v11.3_u1 (bug fixes)&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--  When removing system status entries, please archive them to: --&amp;gt;&lt;br /&gt;
[[Previous messages]]&lt;br /&gt;
&lt;br /&gt;
{|style=&amp;quot;border-spacing: 10px;width: 100%&amp;quot;&lt;br /&gt;
|valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== QuickStart Guides ==&lt;br /&gt;
* [https://docs.alliancecan.ca/wiki/Trillium_Quickstart Trillium Quickstart]&lt;br /&gt;
* [[HPSS | HPSS archival storage]]&lt;br /&gt;
* [[Teach|Teach cluster]]&lt;br /&gt;
* [[FAQ | FAQ (frequently asked questions)]]&lt;br /&gt;
* [[Acknowledging SciNet]]&lt;br /&gt;
| valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== Tutorials, Manuals, etc. ==&lt;br /&gt;
* [https://education.scinet.utoronto.ca SciNet education material]&lt;br /&gt;
* [https://www.youtube.com/c/SciNetHPCattheUniversityofToronto SciNet's YouTube channel]&lt;br /&gt;
* [[Modules for Mist]] &lt;br /&gt;
* [[Commercial software]]&lt;br /&gt;
* [[Burst Buffer]]&lt;br /&gt;
* [[SSH#SSH Keys|SSH keys]]&lt;br /&gt;
* [[SSH Tunneling]]&lt;br /&gt;
* [[Visualization]]&lt;br /&gt;
* [[Running Serial Jobs on Niagara]]&lt;br /&gt;
* [[Jupyter Hub]]&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Rzon</name></author>
	</entry>
	<entry>
		<id>https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7667</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7667"/>
		<updated>2026-04-28T14:36:04Z</updated>

		<summary type="html">&lt;p&gt;Rzon: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;__NOTOC__&lt;br /&gt;
{| style=&amp;quot;border-spacing:10px; width: 95%&amp;quot;&lt;br /&gt;
| style=&amp;quot;padding:1em; padding-top:.1em; border:2px solid #0645ad; background-color:#f6f6f6; border-radius:7px&amp;quot;|&lt;br /&gt;
&lt;br /&gt;
==System Status==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- Use &amp;quot;Up&amp;quot;, &amp;quot;Partial&amp;quot; or &amp;quot;Down&amp;quot;; these are templates. --&amp;gt;&lt;br /&gt;
{|style=&amp;quot;width:100%&amp;quot; &lt;br /&gt;
|{{Up3 | Trillium|https://docs.alliancecan.ca/wiki/Trillium_Quickstart}}&lt;br /&gt;
|{{Up3 | OnDemand|https://docs.alliancecan.ca/wiki/Trillium_Open_OnDemand_Quickstart}}&lt;br /&gt;
|{{Up | Globus |Globus}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Up | HPSS|HPSS}}&lt;br /&gt;
|{{Up | Balam|Balam}}&lt;br /&gt;
|{{Up | S4H | S4H}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Up | Teach|Teach}}&lt;br /&gt;
|{{Up3 | File system|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Storage}}&lt;br /&gt;
|{{Up3 | External Network|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Logging_in}} &lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Thu Apr 23, 2026, 10:00 am:''' The Trillium file system has issues and may be slow on certain nodes. We are still investigating.&lt;br /&gt;
&lt;br /&gt;
'''Tue Apr 14, 2026, 10:30 am:''' The Trillium globus endpoint is working again.&lt;br /&gt;
&lt;br /&gt;
'''Sat Apr 11, 2026, 10:00 pm:''' The Trillium globus endpoint is not operational (it times out). We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''Thu Apr 09, 2026, 10:30 am:''' tri-dm4.scinet.utoronto.ca and robot4.scinet.utoronto.ca are in maintenance. Use 1,2, or 3 instead.&lt;br /&gt;
&lt;br /&gt;
'''Wed Apr 08, 2026, 6:30 pm:''' Software from the CVMFS 'restricted' area is back.&lt;br /&gt;
&lt;br /&gt;
'''Wed Apr 08, 2026, 1:40 pm:''' Software from the CVMFS 'restricted' area is not available on many of the Trillium and Open OnDemand nodes. We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''Mon Apr 06, 2026, 10:00 pm:''' We will have to reschedule the HPSS update. This attempt didn't work as expected&lt;br /&gt;
&lt;br /&gt;
'''Mon Apr 06, 2026, 8:00 pm:''' HPSS scheduled maintenance: update of HPSS to v11.3_u4 and hsi-htar to v11.3_u1 (bug fixes)&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--  When removing system status entries, please archive them to: --&amp;gt;&lt;br /&gt;
[[Previous messages]]&lt;br /&gt;
&lt;br /&gt;
{|style=&amp;quot;border-spacing: 10px;width: 100%&amp;quot;&lt;br /&gt;
|valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== QuickStart Guides ==&lt;br /&gt;
* [https://docs.alliancecan.ca/wiki/Trillium_Quickstart Trillium Quickstart]&lt;br /&gt;
* [[HPSS | HPSS archival storage]]&lt;br /&gt;
* [[Teach|Teach cluster]]&lt;br /&gt;
* [[FAQ | FAQ (frequently asked questions)]]&lt;br /&gt;
* [[Acknowledging SciNet]]&lt;br /&gt;
| valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== Tutorials, Manuals, etc. ==&lt;br /&gt;
* [https://education.scinet.utoronto.ca SciNet education material]&lt;br /&gt;
* [https://www.youtube.com/c/SciNetHPCattheUniversityofToronto SciNet's YouTube channel]&lt;br /&gt;
* [[Modules for Mist]] &lt;br /&gt;
* [[Commercial software]]&lt;br /&gt;
* [[Burst Buffer]]&lt;br /&gt;
* [[SSH#SSH Keys|SSH keys]]&lt;br /&gt;
* [[SSH Tunneling]]&lt;br /&gt;
* [[Visualization]]&lt;br /&gt;
* [[Running Serial Jobs on Niagara]]&lt;br /&gt;
* [[Jupyter Hub]]&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Rzon</name></author>
	</entry>
	<entry>
		<id>https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7664</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7664"/>
		<updated>2026-04-24T00:33:49Z</updated>

		<summary type="html">&lt;p&gt;Rzon: /* System Status */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;__NOTOC__&lt;br /&gt;
{| style=&amp;quot;border-spacing:10px; width: 95%&amp;quot;&lt;br /&gt;
| style=&amp;quot;padding:1em; padding-top:.1em; border:2px solid #0645ad; background-color:#f6f6f6; border-radius:7px&amp;quot;|&lt;br /&gt;
&lt;br /&gt;
==System Status==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- Use &amp;quot;Up&amp;quot;, &amp;quot;Partial&amp;quot; or &amp;quot;Down&amp;quot;; these are templates. --&amp;gt;&lt;br /&gt;
{|style=&amp;quot;width:100%&amp;quot; &lt;br /&gt;
|{{Up3 | Trillium|https://docs.alliancecan.ca/wiki/Trillium_Quickstart}}&lt;br /&gt;
|{{Up3 | OnDemand|https://docs.alliancecan.ca/wiki/Trillium_Open_OnDemand_Quickstart}}&lt;br /&gt;
|{{Up | Globus |Globus}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Up | HPSS|HPSS}}&lt;br /&gt;
|{{Up | Balam|Balam}}&lt;br /&gt;
|{{Up | S4H | S4H}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Up | Teach|Teach}}&lt;br /&gt;
|{{Partial3 | File system|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Storage}}&lt;br /&gt;
|{{Up3 | External Network|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Logging_in}} &lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Thu Apr 23, 2026, 10:00 am:''' The Trillium file system has issues and may be slow on certain nodes. We are still investigating.&lt;br /&gt;
&lt;br /&gt;
'''Tue Apr 14, 2026, 10:30 am:''' The Trillium globus endpoint is working again.&lt;br /&gt;
&lt;br /&gt;
'''Sat Apr 11, 2026, 10:00 pm:''' The Trillium globus endpoint is not operational (it times out). We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''Thu Apr 09, 2026, 10:30 am:''' tri-dm4.scinet.utoronto.ca and robot4.scinet.utoronto.ca are in maintenance. Use 1,2, or 3 instead.&lt;br /&gt;
&lt;br /&gt;
'''Wed Apr 08, 2026, 6:30 pm:''' Software from the CVMFS 'restricted' area is back.&lt;br /&gt;
&lt;br /&gt;
'''Wed Apr 08, 2026, 1:40 pm:''' Software from the CVMFS 'restricted' area is not available on many of the Trillium and Open OnDemand nodes. We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''Mon Apr 06, 2026, 10:00 pm:''' We will have to reschedule the HPSS update. This attempt didn't work as expected&lt;br /&gt;
&lt;br /&gt;
'''Mon Apr 06, 2026, 8:00 pm:''' HPSS scheduled maintenance: update of HPSS to v11.3_u4 and hsi-htar to v11.3_u1 (bug fixes)&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--  When removing system status entries, please archive them to: --&amp;gt;&lt;br /&gt;
[[Previous messages]]&lt;br /&gt;
&lt;br /&gt;
{|style=&amp;quot;border-spacing: 10px;width: 100%&amp;quot;&lt;br /&gt;
|valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== QuickStart Guides ==&lt;br /&gt;
* [https://docs.alliancecan.ca/wiki/Trillium_Quickstart Trillium Quickstart]&lt;br /&gt;
* [[HPSS | HPSS archival storage]]&lt;br /&gt;
* [[Teach|Teach cluster]]&lt;br /&gt;
* [[FAQ | FAQ (frequently asked questions)]]&lt;br /&gt;
* [[Acknowledging SciNet]]&lt;br /&gt;
| valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== Tutorials, Manuals, etc. ==&lt;br /&gt;
* [https://education.scinet.utoronto.ca SciNet education material]&lt;br /&gt;
* [https://www.youtube.com/c/SciNetHPCattheUniversityofToronto SciNet's YouTube channel]&lt;br /&gt;
* [[Modules for Mist]] &lt;br /&gt;
* [[Commercial software]]&lt;br /&gt;
* [[Burst Buffer]]&lt;br /&gt;
* [[SSH#SSH Keys|SSH keys]]&lt;br /&gt;
* [[SSH Tunneling]]&lt;br /&gt;
* [[Visualization]]&lt;br /&gt;
* [[Running Serial Jobs on Niagara]]&lt;br /&gt;
* [[Jupyter Hub]]&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Rzon</name></author>
	</entry>
	<entry>
		<id>https://docs.scinet.utoronto.ca/index.php?title=Niagara_Quickstart&amp;diff=7661</id>
		<title>Niagara Quickstart</title>
		<link rel="alternate" type="text/html" href="https://docs.scinet.utoronto.ca/index.php?title=Niagara_Quickstart&amp;diff=7661"/>
		<updated>2026-04-15T19:17:38Z</updated>

		<summary type="html">&lt;p&gt;Rzon: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Infobox Computer&lt;br /&gt;
|image=[[Image:Niagara.jpg|center|300px|thumb]]&lt;br /&gt;
|name=Niagara&lt;br /&gt;
|installed=Jan 2018/March 2020&lt;br /&gt;
|operatingsystem= CentOS 7.9 &lt;br /&gt;
|loginnode= niagara.scinet.utoronto.ca&lt;br /&gt;
|nnodes=  2,024 nodes (80,960 cores)&lt;br /&gt;
|rampernode=188 GiB / 202 GB  &lt;br /&gt;
|corespernode=40 (80 hyperthreads)&lt;br /&gt;
|interconnect=Mellanox Dragonfly+&lt;br /&gt;
|vendorcompilers= icc (C) ifort (fortran) icpc (C++)&lt;br /&gt;
|queuetype=Slurm&lt;br /&gt;
}}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
'''&amp;lt;font color=red&amp;gt;The Niagara cluster was &amp;lt;b&amp;gt;decommissioned&amp;lt;/b&amp;gt; on September 30, 2025. &amp;lt;br /&amp;gt;Its users should move to the  [https://docs.alliancecan.ca/wiki/Trillium_Quickstart Trillium]. &amp;lt;br /&amp;gt;None of the instructions below will work on Trillium.&amp;lt;/font&amp;gt;'''&lt;br /&gt;
&lt;br /&gt;
=Specifications=&lt;br /&gt;
&lt;br /&gt;
The Niagara cluster is a large cluster of 2,024 Lenovo SD530 servers each with 40 Intel &amp;quot;Skylake&amp;quot; cores at 2.4 GHz (1548 nodes) or 40 Intel &amp;quot;CascadeLake&amp;quot; cores at 2.5 GHz (476 nodes). &lt;br /&gt;
The peak performance of the cluster is about 3.6 PFlops (6.25 PFlops theoretical).  It was the 53rd fastest supercomputer on the [https://www.top500.org/list/2018/06/?page=1 TOP500 list of June 2018], and is at number 241 on the [https://www.top500.org/lists/top500/list/2024/06/?page=3 current list (June 2024)]. &lt;br /&gt;
&lt;br /&gt;
Each node of the cluster has 188 GiB / 202 GB RAM per node (at least 4 GiB/core for user jobs and roughly 170 GiB/node at most).  Being designed for large parallel workloads, it has a fast interconnect consisting of EDR InfiniBand in a Dragonfly+ topology with Adaptive Routing.  The compute nodes are accessed through a queueing system that allows jobs with a minimum of 15 minutes and a maximum of 24 hours and favours large jobs.&lt;br /&gt;
&lt;br /&gt;
* See the [https://www.youtube.com/watch?v=l-E2CFGh0BE&amp;amp;feature=youtu.be  &amp;quot;Intro to Niagara&amp;quot;] recording&lt;br /&gt;
&lt;br /&gt;
More detailed hardware characteristics of the Niagara supercomputer can be found [https://docs.alliancecan.ca/wiki/Niagara on this page].&lt;br /&gt;
&lt;br /&gt;
Note: Documentation about the &amp;quot;GPU expansion to Niagara&amp;quot; called &amp;quot;Mist&amp;quot; can be found on [[Mist | its own page]].&lt;br /&gt;
&lt;br /&gt;
= Getting started on Niagara =&lt;br /&gt;
&lt;br /&gt;
Access to Niagara is not enabled automatically for everyone with an account with the {{DigitalResearchAllianceOfCanada}}, but anyone with an active Alliance account can get their access enabled.&lt;br /&gt;
 &lt;br /&gt;
If you have an active Alliance account but you do not have access to Niagara yet (e.g. because you are new to SciNet or belong to a group whose primary PI does not have an allocation as granted in the annual [https://alliancecan.ca/en/services/advanced-research-computing/research-portal/accessing-resources/resource-allocation-competitions {{Alliance}} RAC]), go to the [https://ccdb.alliancecan.ca/services/opt_in opt-in page on the CCDB site].  After clicking the &amp;quot;Join&amp;quot; button, it usually takes only one or two business days for access to be granted.  &lt;br /&gt;
&lt;br /&gt;
Please read this document carefully.  The [https://docs.scinet.utoronto.ca/index.php/FAQ FAQ] is also a useful resource.  If at any time you require assistance, or if something is unclear, please do not hesitate to [mailto:support@scinet.utoronto.ca contact us].&lt;br /&gt;
&lt;br /&gt;
== Logging in ==&lt;br /&gt;
&lt;br /&gt;
There are two ways to access Niagara:&lt;br /&gt;
&lt;br /&gt;
# Via your browser with Open OnDemand. This is recommended for users who are not familiar with Linux or the command line. Please see our [https://docs.scinet.utoronto.ca/index.php/Open_OnDemand_Quickstart quickstart guide] for more instructions on how to use Open OnDemand.&lt;br /&gt;
# Terminal access with ssh. Please read the following instructions.&lt;br /&gt;
        &lt;br /&gt;
Niagara runs CentOS 7, which is a type of Linux.  You will need to be familiar with Linux systems to work on Niagara.  If you are not it will be worth your time to review our [https://support.scinet.utoronto.ca/education/browse.php?category=-1&amp;amp;search=scmp101&amp;amp;include=all&amp;amp;filter=Filter Introduction to Linux Shell] class.&lt;br /&gt;
&lt;br /&gt;
As with all SciNet and {{Alliance}} compute systems, access to Niagara is done via [[SSH]] (secure shell) only. As of January 22 2022, authentication is only allowed via SSH keys. [https://docs.alliancecan.ca/wiki/SSH_Keys Please refer to this page] to generate your SSH key pair and make sure you use them securely.&lt;br /&gt;
 &lt;br /&gt;
Open a terminal window (e.g. Connecting with [https://docs.alliancecan.ca/wiki/Connecting_with_PuTTY PuTTY] on Windows or Connecting with [https://docs.alliancecan.ca/wiki/Connecting_with_MobaXTerm MobaXTerm]), then SSH into the Niagara login nodes with your {{Alliance}} credentials:&lt;br /&gt;
&lt;br /&gt;
 $ ssh -i /path/to/ssh_private_key -Y MYALLIANCEUSERNAME@niagara.scinet.utoronto.ca&lt;br /&gt;
&lt;br /&gt;
or&lt;br /&gt;
&lt;br /&gt;
 $ ssh -i /path/to/ssh_private_key -Y MYALLIANCEUSERNAME@niagara.computecanada.ca&lt;br /&gt;
&lt;br /&gt;
The first time you login to Niagara, please make sure you are actually accessing Niagara by checking if the login node ssh host key fingerprint matches [[SSH_Changes_in_May_2019 | (See here how)]]. This check prevents you from falling victim of [https://en.wikipedia.org/wiki/Man-in-the-middle_attack man-in-the-middle attacks.]&lt;br /&gt;
&lt;br /&gt;
* The Niagara login nodes are where you develop, edit, compile, prepare and submit jobs.&lt;br /&gt;
* These login nodes are not part of the Niagara compute cluster, but have the same architecture, operating system, and software stack.&lt;br /&gt;
* The optional &amp;lt;code&amp;gt;-Y&amp;lt;/code&amp;gt; is needed to open windows from the Niagara command-line onto your local X server.&lt;br /&gt;
* You can only connect 4 times in a 2-minute window to the login nodes. &lt;br /&gt;
* To run on Niagara's compute nodes, you must [[#Submitting_jobs | submit a batch job]].&lt;br /&gt;
&lt;br /&gt;
If you cannot log in, be sure to first check the [https://docs.scinet.utoronto.ca System Status] on this site's front page.&lt;br /&gt;
&lt;br /&gt;
== Your various directories ==&lt;br /&gt;
&lt;br /&gt;
By virtue of your access to Niagara you are granted storage space on the system.  There are several directories available to you, each indicated by an associated environment variable.&lt;br /&gt;
&lt;br /&gt;
=== home and scratch ===&lt;br /&gt;
&lt;br /&gt;
You have a home and scratch directory on the system, the paths to which are stored in the environment variables $HOME and $SCRATCH. The locations are of the form&lt;br /&gt;
&lt;br /&gt;
 $HOME=/home/g/groupname/myallianceusername&lt;br /&gt;
 $SCRATCH=/scratch/g/groupname/myallianceusername&lt;br /&gt;
&lt;br /&gt;
where groupname is the name of your PI's group, and myallianceusername is your {{Alliance}} username.  For example:&lt;br /&gt;
&lt;br /&gt;
  nia-login07:~$ pwd&lt;br /&gt;
  /home/s/scinet/rzon&lt;br /&gt;
  nia-login07:~$ cd $SCRATCH&lt;br /&gt;
  nia-login07:rzon$ pwd&lt;br /&gt;
  /scratch/s/scinet/rzon&lt;br /&gt;
&lt;br /&gt;
NOTE: home is read-only on compute nodes.&lt;br /&gt;
&lt;br /&gt;
=== project and archive/nearline ===&lt;br /&gt;
&lt;br /&gt;
Users from groups with [https://www.alliancecan.ca/research-portal/accessing-resources/resource-allocation-competitions RAC storage allocation] will also have a project directory and possible an archive (a.k.a. &amp;quot;nearline&amp;quot;) directory, the paths to which are stored in the environment variables $PROJECT and $ARCHIVE. They follow the naming convention:&lt;br /&gt;
&lt;br /&gt;
 $PROJECT=/project/g/groupname/myallianceusername&lt;br /&gt;
 $ARCHIVE=/archive/g/groupname/myallianceusername&lt;br /&gt;
&lt;br /&gt;
NOTE: Currently archive space is available only via [[HPSS]], and is not accessible on the Niagara login, compute, or datamover nodes.&lt;br /&gt;
&lt;br /&gt;
'''''IMPORTANT: Future-proof your scripts'''''&lt;br /&gt;
&lt;br /&gt;
When writing your scripts, use the environment variables (&amp;lt;tt&amp;gt;$HOME&amp;lt;/tt&amp;gt;, &amp;lt;tt&amp;gt;$SCRATCH&amp;lt;/tt&amp;gt;, &amp;lt;tt&amp;gt;$PROJECT&amp;lt;/tt&amp;gt;, &amp;lt;tt&amp;gt;$ARCHIVE&amp;lt;/tt&amp;gt;) instead of the actual paths!  The paths may change in the future.&lt;br /&gt;
&lt;br /&gt;
=== Storage and quotas ===&lt;br /&gt;
&lt;br /&gt;
You should familiarize yourself with the [[Data_Management#Purpose_of_each_file_system | various file systems]], what purpose they serve, and how to properly use them.  This table summarizes the various file systems.  See the [[Data_Management | Data Management]] page for more details.&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
! location&lt;br /&gt;
!colspan=&amp;quot;2&amp;quot;| quota&lt;br /&gt;
!align=&amp;quot;right&amp;quot;| block size&lt;br /&gt;
! expiration time&lt;br /&gt;
! backed up&lt;br /&gt;
! on login nodes&lt;br /&gt;
! on compute nodes&lt;br /&gt;
|-&lt;br /&gt;
| $HOME&lt;br /&gt;
|colspan=&amp;quot;2&amp;quot;| 100 GB / 250,000 files per user&lt;br /&gt;
|align=&amp;quot;right&amp;quot;| 1 MB&lt;br /&gt;
| &lt;br /&gt;
| yes&lt;br /&gt;
| yes&lt;br /&gt;
| read-only&lt;br /&gt;
|-&lt;br /&gt;
|rowspan=&amp;quot;2&amp;quot;| $SCRATCH&lt;br /&gt;
|colspan=&amp;quot;2&amp;quot;| 25 TB / 6,000,000 file per user&lt;br /&gt;
|align=&amp;quot;right&amp;quot; rowspan=&amp;quot;2&amp;quot; | 16 MB&lt;br /&gt;
|rowspan=&amp;quot;2&amp;quot;| 2 months&lt;br /&gt;
|rowspan=&amp;quot;2&amp;quot;| no&lt;br /&gt;
|rowspan=&amp;quot;2&amp;quot;| yes&lt;br /&gt;
|rowspan=&amp;quot;2&amp;quot;| yes&lt;br /&gt;
|-&lt;br /&gt;
|align=&amp;quot;right&amp;quot;|50-500TB per group&lt;br /&gt;
|align=&amp;quot;right&amp;quot;|[[Data_Management#Quotas_and_purging | depending on group size]]&lt;br /&gt;
|-&lt;br /&gt;
| $PROJECT&lt;br /&gt;
|colspan=&amp;quot;2&amp;quot;| by group allocation&lt;br /&gt;
|align=&amp;quot;right&amp;quot;| 16 MB&lt;br /&gt;
| &lt;br /&gt;
| yes&lt;br /&gt;
| yes&lt;br /&gt;
| yes&lt;br /&gt;
|-&lt;br /&gt;
| $ARCHIVE&lt;br /&gt;
|colspan=&amp;quot;2&amp;quot;| by group (nearline) allocation&lt;br /&gt;
|align=&amp;quot;right&amp;quot;| &lt;br /&gt;
|&lt;br /&gt;
| dual-copy&lt;br /&gt;
| no&lt;br /&gt;
| no&lt;br /&gt;
|-&lt;br /&gt;
| $BBUFFER&lt;br /&gt;
|colspan=&amp;quot;2&amp;quot;| 10 TB per user&lt;br /&gt;
|align=&amp;quot;right&amp;quot;| 1 MB&lt;br /&gt;
| very short&lt;br /&gt;
| no&lt;br /&gt;
| yes&lt;br /&gt;
| yes&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== Moving data to Niagara ==&lt;br /&gt;
&lt;br /&gt;
If you need to move data to Niagara for analysis, or when you need to move data off of Niagara, use the following guidelines:&lt;br /&gt;
* If your data is less than 10GB, move the data using the login nodes.&lt;br /&gt;
* If your data is greater than 10GB, move the data using the datamover nodes nia-datamover1.scinet.utoronto.ca and nia-datamover2.scinet.utoronto.ca .&lt;br /&gt;
&lt;br /&gt;
Details of how to use the datamover nodes can be found on the [[Data_Management#Moving_data | Data Management ]] page.&lt;br /&gt;
&lt;br /&gt;
= Loading software modules =&lt;br /&gt;
&lt;br /&gt;
You have two options for running code on Niagara: use existing software, or [[Niagara_Quickstart#Compiling_on_Niagara:_Example | compile your own]].  This section focuses on the former.&lt;br /&gt;
&lt;br /&gt;
Other than essentials, all installed software is made available [[Using_modules | using module commands]]. These modules set environment variables (PATH, etc.), allowing multiple, conflicting versions of a given package to be available.  A detailed explanation of the module system can be [[Using_modules | found on the modules page]].&lt;br /&gt;
&lt;br /&gt;
Common module subcommands are:&lt;br /&gt;
&lt;br /&gt;
* &amp;lt;code&amp;gt;module load &amp;lt;module-name&amp;gt;&amp;lt;/code&amp;gt;: load the default version of a particular software.&lt;br /&gt;
* &amp;lt;code&amp;gt;module load &amp;lt;module-name&amp;gt;/&amp;lt;module-version&amp;gt;&amp;lt;/code&amp;gt;: load a specific version of a particular software.&lt;br /&gt;
* &amp;lt;code&amp;gt;module purge&amp;lt;/code&amp;gt;: unload all currently loaded modules.&lt;br /&gt;
* &amp;lt;code&amp;gt;module spider&amp;lt;/code&amp;gt; (or &amp;lt;code&amp;gt;module spider &amp;lt;module-name&amp;gt;&amp;lt;/code&amp;gt;): list available software packages.&lt;br /&gt;
* &amp;lt;code&amp;gt;module avail&amp;lt;/code&amp;gt;: list loadable software packages.&lt;br /&gt;
* &amp;lt;code&amp;gt;module list&amp;lt;/code&amp;gt;: list loaded modules.&lt;br /&gt;
&lt;br /&gt;
Along with modifying common environment variables, such as PATH, and LD_LIBRARY_PATH, these modules also create a SCINET_MODULENAME_ROOT environment variable, which can be used to access commonly needed software directories, such as /include and /lib.&lt;br /&gt;
&lt;br /&gt;
There are handy abbreviations for the module commands. &amp;lt;code&amp;gt;ml&amp;lt;/code&amp;gt; is the same as &amp;lt;code&amp;gt;module list&amp;lt;/code&amp;gt;, and &amp;lt;code&amp;gt;ml &amp;lt;module-name&amp;gt;&amp;lt;/code&amp;gt; is the same as &amp;lt;code&amp;gt;module load &amp;lt;module-name&amp;gt;&amp;lt;/code&amp;gt;.&lt;br /&gt;
== Software stacks: NiaEnv and CCEnv ==&lt;br /&gt;
&lt;br /&gt;
On Niagara, there are two available software stacks:&lt;br /&gt;
&lt;br /&gt;
=== NiaEnv ===&lt;br /&gt;
&lt;br /&gt;
A [https://docs.scinet.utoronto.ca/index.php/Modules_specific_to_Niagara Niagara software stack] tuned and compiled for this machine. This stack is available by default, but if not, can be reloaded with&lt;br /&gt;
&amp;lt;pre&amp;gt;module load NiaEnv&amp;lt;/pre&amp;gt;&lt;br /&gt;
This loads the default (set of modules), which is currently the 2019b epoch. Before September 1, the default was NiaEnv/2018a.  Since May 2023, a newer stack, NiaEnv/2022a, is available as well.&lt;br /&gt;
&lt;br /&gt;
Existing users are expected to continue using the 2019b stack, but to make old job scripts or older software installations in your home directory work, you may need to use&lt;br /&gt;
&amp;lt;pre&amp;gt;module load NiaEnv/2018a&amp;lt;/pre&amp;gt;&lt;br /&gt;
For new projects, we advice users to start with the 2022a stack:&lt;br /&gt;
&amp;lt;pre&amp;gt;module load NiaEnv/2022a&amp;lt;/pre&amp;gt;&lt;br /&gt;
You can override the system default for the epoch version by creating a file called &amp;lt;b&amp;gt;&amp;lt;tt&amp;gt;.modulerc&amp;lt;/tt&amp;gt;&amp;lt;/b&amp;gt; in your home directory with the line &amp;lt;b&amp;gt;&amp;lt;tt&amp;gt;module-version NiaEnv/VERSION default&amp;lt;/tt&amp;gt;&amp;lt;/b&amp;gt;, e.g. like so:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
echo &amp;quot;module-version NiaEnv/2022a default&amp;quot; &amp;gt; $HOME/.modulerc&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
After this, subsequent logins and jobs will use the 2022a stack even when the system default is different.&lt;br /&gt;
&amp;lt;p&amp;gt;Similarly, you can make an older epoch your personal default, like so&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
echo &amp;quot;module-version NiaEnv/2018a default&amp;quot; &amp;gt; $HOME/.modulerc&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
No modules are loaded by default on Niagara except NiaEnv.&lt;br /&gt;
&lt;br /&gt;
=== CCEnv ===&lt;br /&gt;
&lt;br /&gt;
The same  [https://docs.alliancecan.ca/wiki/Modules software stack available on {{Alliance}}'s General Purpose clusters]  too, with:&lt;br /&gt;
&amp;lt;pre&amp;gt;module load CCEnv&amp;lt;/pre&amp;gt;&lt;br /&gt;
Or, if you want the same default modules loaded as on Béluga, then do&lt;br /&gt;
&amp;lt;pre&amp;gt;module load CCEnv StdEnv&amp;lt;/pre&amp;gt;&lt;br /&gt;
or, if you want the same default modules loaded as on Cedar and Graham, do&lt;br /&gt;
&amp;lt;pre&amp;gt;module load CCEnv arch/avx2 StdEnv/2020&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Tips for loading software ==&lt;br /&gt;
&lt;br /&gt;
* We advise '''''against''''' loading modules in your .bashrc.  This can lead to very confusing behaviour under certain circumstances.  Our guidelines for .bashrc files can be found [[bashrc guidelines|here]].&lt;br /&gt;
* Instead, load modules by hand when needed, or by sourcing a separate script.&lt;br /&gt;
* Load run-specific modules inside your job submission script.&lt;br /&gt;
* Short names give default versions; e.g. &amp;lt;code&amp;gt;intel&amp;lt;/code&amp;gt; → &amp;lt;code&amp;gt;intel/2018.2&amp;lt;/code&amp;gt;. It is usually better to be explicit about the versions, for future reproducibility.&lt;br /&gt;
* Modules often require other modules to be loaded first.  Solve these dependencies by using [[Using_modules#Module_spider | &amp;lt;code&amp;gt;module spider&amp;lt;/code&amp;gt;]].&lt;br /&gt;
&lt;br /&gt;
= Available compilers and interpreters =&lt;br /&gt;
&lt;br /&gt;
* For most compiled software, one should use the Intel compilers (&amp;lt;tt&amp;gt;icc&amp;lt;/tt&amp;gt; for C, &amp;lt;tt&amp;gt;icpc&amp;lt;/tt&amp;gt; for C++, and &amp;lt;tt&amp;gt;ifort&amp;lt;/tt&amp;gt; for Fortran). Loading an &amp;lt;tt&amp;gt;intel&amp;lt;/tt&amp;gt; module makes these available. &lt;br /&gt;
* The GNU compiler suite (&amp;lt;tt&amp;gt;gcc, g++, gfortran&amp;lt;/tt&amp;gt;) is also available, if you load one of the &amp;lt;tt&amp;gt;gcc&amp;lt;/tt&amp;gt; modules.&lt;br /&gt;
* To compile mpi code, you must additionally load an &amp;lt;tt&amp;gt;openmpi&amp;lt;/tt&amp;gt; or &amp;lt;tt&amp;gt;intelmpi&amp;lt;/tt&amp;gt; module.&lt;br /&gt;
* Open source interpreted, interactive software is also available:&lt;br /&gt;
** [[Python]]&lt;br /&gt;
** [[R]]&lt;br /&gt;
** Julia&lt;br /&gt;
** [[Octave]]&lt;br /&gt;
  &lt;br /&gt;
Please visit the corresponding page for details on using these tools.  For information on running MATLAB applications on Niagara, visit [[MATLAB| this page]].&lt;br /&gt;
&lt;br /&gt;
= Using Commercial Software =&lt;br /&gt;
&lt;br /&gt;
May I use commercial software on Niagara?&lt;br /&gt;
* Possibly, but you have to bring your own license for it.  You can connect to an external license server using [[SSH_Tunneling | ssh tunneling]].&lt;br /&gt;
* SciNet and {{the Alliance}} have an extremely large and broad user base of thousands of users, so we cannot provide licenses for everyone's favorite software.&lt;br /&gt;
* Thus, the only freely available commercial software installed on Niagara is software that can benefit everyone: Compilers, math libraries and debuggers.&lt;br /&gt;
* That means no [[MATLAB]], Gaussian, IDL, &lt;br /&gt;
* Open source alternatives like Octave, [[Python]], and [[R]] are available.&lt;br /&gt;
* We are happy to help you to install commercial software for which you have a license.&lt;br /&gt;
* In some cases, if you have a license, you can use software in the {{Alliance}} stack.&lt;br /&gt;
The list of commercial software which is installed on Niagara, for which you will need a license to use, can be found on the [[Commercial_software | commercial software page]].&lt;br /&gt;
&lt;br /&gt;
= Compiling on Niagara: Example =&lt;br /&gt;
&lt;br /&gt;
Suppose one wants to compile an application from two c source files, appl.c and module.c, which use the Math Kernel Library. This is an example of how this would be done:&lt;br /&gt;
&amp;lt;source lang=&amp;quot;bash&amp;quot;&amp;gt;&lt;br /&gt;
nia-login07:~$ module load NiaEnv/2019b&lt;br /&gt;
nia-login07:~$ module list&lt;br /&gt;
Currently Loaded Modules:&lt;br /&gt;
  1) NiaEnv/2019b (S)&lt;br /&gt;
  Where:&lt;br /&gt;
   S:  Module is Sticky, requires --force to unload or purge&lt;br /&gt;
&lt;br /&gt;
nia-login07:~$ module load intel/2019u4&lt;br /&gt;
&lt;br /&gt;
nia-login07:~$ ls&lt;br /&gt;
appl.c module.c&lt;br /&gt;
&lt;br /&gt;
nia-login07:~$ icc -c -O3 -xHost -o appl.o appl.c&lt;br /&gt;
nia-login07:~$ icc -c -O3 -xHost -o module.o module.c&lt;br /&gt;
nia-login07:~$ icc  -o appl module.o appl.o -mkl&lt;br /&gt;
&lt;br /&gt;
nia-login07:~$ ./appl&lt;br /&gt;
&amp;lt;/source&amp;gt;&lt;br /&gt;
Note:&lt;br /&gt;
* The optimization flags -O3 -xHost allow the Intel compiler to use instructions specific to the architecture CPU that is present (instead of for more generic x86_64 CPUs).&lt;br /&gt;
* Linking with the Intel Math Kernel Library (MKL) is easy when using the intel compiler, it just requires the -mkl flags.&lt;br /&gt;
* If compiling with gcc, the optimization flags would be -O3 -march=native. For the way to link with the MKL, it is suggested to use the [https://software.intel.com/en-us/articles/intel-mkl-link-line-advisor MKL link line advisor].&lt;br /&gt;
&lt;br /&gt;
= Testing and Debugging =&lt;br /&gt;
&lt;br /&gt;
You really should test your code before you submit it to the cluster to know if your code is correct and what kind of resources you need.&lt;br /&gt;
* Small test jobs can be run on the login nodes.  Rule of thumb: tests should run no more than a couple of minutes, taking at most about 1-2GB of memory, and use no more than a couple of cores.&lt;br /&gt;
* You can run the [[Parallel Debugging with DDT|DDT]] debugger on the login nodes after &amp;lt;code&amp;gt;module load ddt&amp;lt;/code&amp;gt;.&lt;br /&gt;
* Short tests that do not fit on a login node, or for which you need a dedicated node, request an interactive debug job with the debug command:&lt;br /&gt;
 nia-login07:~$ debugjob --clean N&lt;br /&gt;
where N is the number of nodes, If N=1, this gives an interactive session one 1 hour, when N=4 (the maximum), it gives you 22 minutes.  The &amp;lt;tt&amp;gt;--clean&amp;lt;/tt&amp;gt; argument is optional but recommended as it will start the session without any modules loaded, thus mimicking more closely what happens when you submit a job script.&lt;br /&gt;
&lt;br /&gt;
Finally, if your debugjob process takes more than 1 hour, you can request an interactive job from the regular queue using the salloc command.  Note, however, that this may take some time to run, since it will be part of the regular queue, and will be run when the scheduler decides.&lt;br /&gt;
 nia-login07:~$ salloc --nodes N --time=M:00:00 --x11&lt;br /&gt;
where N is again the number of nodes, and M is the number of hours you wish the job to run.&lt;br /&gt;
The &amp;lt;tt&amp;gt;--x11&amp;lt;/tt&amp;gt; is required if you need to use graphics while testing your code through salloc, e.g. when using a debugger such as [[Parallel Debugging with DDT|DDT]] or DDD, See the [[Testing_With_Graphics | Testing with graphics]] page for the options in that case.&lt;br /&gt;
&lt;br /&gt;
= Submitting jobs =&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- == Progressive approach to run jobs on niagara == --&amp;gt;&lt;br /&gt;
&amp;lt;!-- We would like to emphasize the need for users to adopt a more progressive and explicit approach for testing, running and scaling up of jobs on niagara. [[Progressive_Approach | '''Here is a set of steps we suggest that you follow.''']] --&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Once you have compiled and tested your code or workflow on the Niagara login nodes, and confirmed that it behaves correctly, you are ready to submit jobs to the cluster.  Your jobs will run on some of Niagara's 1548 compute nodes.  When and where your job runs is determined by the scheduler.&lt;br /&gt;
&lt;br /&gt;
Niagara uses SLURM as its job scheduler.  More-advanced details of how to interact with the scheduler can be found on the [[Slurm | Slurm page]].&lt;br /&gt;
&lt;br /&gt;
You submit jobs from a login node by passing a script to the sbatch command:&lt;br /&gt;
&lt;br /&gt;
 nia-login07:scratch$ sbatch jobscript.sh&lt;br /&gt;
&lt;br /&gt;
This puts the job in the queue. It will run on the compute nodes in due course.  Note that you must submit your job from a login node.  You cannot submit jobs from the datamover nodes.&lt;br /&gt;
&lt;br /&gt;
In most cases, you should not submit from your $HOME directory, but rather, from your $SCRATCH directory, so that the output of your compute job can be written out (as mentioned above, $HOME is read-only on the compute nodes).&lt;br /&gt;
&lt;br /&gt;
Jobs will run under your group's RRG allocation, or, if the your group has none, under a RAS allocation (previously called `default' allocation).&lt;br /&gt;
&lt;br /&gt;
Some example job scripts can be found below.&lt;br /&gt;
&lt;br /&gt;
Keep in mind:&lt;br /&gt;
* Scheduling is by node, so in multiples of 40 cores.&lt;br /&gt;
* Your job's maximum walltime is 24 hours. &lt;br /&gt;
* Jobs must write their output to your scratch or project directory (home is read-only on compute nodes).&lt;br /&gt;
* Compute nodes have no internet access.&lt;br /&gt;
* Your job script will not remember the modules you have loaded, so it needs to contain &amp;quot;module load&amp;quot; commands of all the required modules (see examples below). &lt;br /&gt;
* [[Data_Management#Moving_data | Move your data]] to Niagara before you submit your job.&lt;br /&gt;
&lt;br /&gt;
== Scheduling by Node ==&lt;br /&gt;
&lt;br /&gt;
On many systems that use SLURM, the scheduler will deduce from the specifications of the number of tasks and the number of cpus-per-node what resources should be allocated.  On Niagara things are a bit different.&lt;br /&gt;
* All job resource requests on Niagara are scheduled as a multiple of '''nodes'''.&lt;br /&gt;
* The nodes that your jobs run on are exclusively yours, for as long as the job is running on them.&lt;br /&gt;
** No other users are running anything on them.&lt;br /&gt;
** You can [[SSH]] into them to see how things are going.&lt;br /&gt;
* Whatever your requests to the scheduler, it will always be translated into a multiple of nodes allocated to your job.&lt;br /&gt;
* Memory requests to the scheduler are of no use. Your job always gets N x 202GB of RAM, where N is the number of nodes and 202GB is the amount of memory on the node.&lt;br /&gt;
* If you run serial jobs you must still use all 40 cores on the node.  Visit the [[Running_Serial_Jobs_on_Niagara | serial jobs]] page for examples of how to do this.&lt;br /&gt;
* Since there are 40 cores per node, your job should use N x 40 cores. If you do not, we will contact you to help you optimize your workflow.  Or you can [mailto:support@scinet.utoronto.ca contact us] to get assistance.&lt;br /&gt;
&lt;br /&gt;
== Limits ==&lt;br /&gt;
&lt;br /&gt;
There are limits to the size and duration of your jobs, the number of jobs you can run and the number of jobs you can have queued.  It matters whether a user is part of a group with a [https://www.alliancecan.ca/research-portal/accessing-resources/resource-allocation-competitions/ Resources for Research Group allocation] or not. It also matters in which 'partition' the job runs. 'Partitions' are SLURM-speak for use cases.  You specify the partition with the &amp;lt;tt&amp;gt;-p&amp;lt;/tt&amp;gt; parameter to &amp;lt;tt&amp;gt;sbatch&amp;lt;/tt&amp;gt; or &amp;lt;tt&amp;gt;salloc&amp;lt;/tt&amp;gt;, but if you do not specify one, your job will run in the &amp;lt;tt&amp;gt;compute&amp;lt;/tt&amp;gt; partition, which is the most common case. &lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!Usage&lt;br /&gt;
!Partition&lt;br /&gt;
!Limit on Running jobs&lt;br /&gt;
!Limit on Submitted jobs (incl. running)&lt;br /&gt;
!Min. size of jobs&lt;br /&gt;
!Max. size of jobs&lt;br /&gt;
!Min. walltime&lt;br /&gt;
!Max. walltime &lt;br /&gt;
|-&lt;br /&gt;
|Compute jobs ||compute || 50 || 1000 || 1 node (40&amp;amp;nbsp;cores) || default:&amp;amp;nbsp;20&amp;amp;nbsp;nodes&amp;amp;nbsp;(800&amp;amp;nbsp;cores) &amp;lt;br&amp;gt; with&amp;amp;nbsp;allocation:&amp;amp;nbsp;1000&amp;amp;nbsp;nodes&amp;amp;nbsp;(40000&amp;amp;nbsp;cores)|| 15 minutes || 24 hours&lt;br /&gt;
|-&lt;br /&gt;
|Testing or troubleshooting || debug || 1 || 1 || 1 node (40&amp;amp;nbsp;cores) || 4 nodes (160 cores)|| N/A || 1 hour&lt;br /&gt;
|-&lt;br /&gt;
|Archiving or retrieving data in [[HPSS]]|| archivelong || 2 per user (5 in total) || 10 per user || N/A || N/A|| 15 minutes || 72 hours&lt;br /&gt;
|-&lt;br /&gt;
|Inspecting archived data, small archival actions in [[HPSS]] || archiveshort vfsshort || 2 per user|| 10 per user || N/A || N/A || 15 minutes || 1 hour&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
Even if you respect these limits, your jobs will still have to wait in the queue.  The waiting time depends on many factors such as your group's allocation amount, how much allocation has been used in the recent past, the number of requested nodes and walltime, and how many other jobs are waiting in the queue.&lt;br /&gt;
&lt;br /&gt;
== File Input/Output Tips ==&lt;br /&gt;
&lt;br /&gt;
It is important to understand the file systems, so as to perform your file I/O (Input/Output) responsibly.  Refer to the [[Data_Management | Data Management]] page for details about the file systems.&lt;br /&gt;
* Your files can be seen on all Niagara login and compute nodes.&lt;br /&gt;
* $HOME, $SCRATCH, and $PROJECT all use the parallel file system called GPFS.&lt;br /&gt;
* GPFS is a high-performance file system which provides rapid reads and writes to large data sets in parallel from many nodes.&lt;br /&gt;
* Accessing data sets which consist of many, small files leads to poor performance on GPFS.&lt;br /&gt;
* Avoid reading and writing lots of small amounts of data to disk.  Many small files on the system waste space and are slower to access, read and write.  If you must write many small files, use [[User_Ramdisk | ramdisk]].&lt;br /&gt;
* Write data out in a binary format. This is faster and takes less space.&lt;br /&gt;
* The [[Burst Buffer]] is another option for I/O heavy-jobs and for speeding up [[Checkpoints|checkpoints]].&lt;br /&gt;
&lt;br /&gt;
== Example submission script (MPI) ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;source lang=&amp;quot;bash&amp;quot;&amp;gt;#!/bin/bash &lt;br /&gt;
#SBATCH --nodes=2&lt;br /&gt;
#SBATCH --ntasks-per-node=40&lt;br /&gt;
#SBATCH --time=1:00:00&lt;br /&gt;
#SBATCH --job-name=mpi_job&lt;br /&gt;
#SBATCH --output=mpi_output_%j.txt&lt;br /&gt;
#SBATCH --mail-type=FAIL&lt;br /&gt;
&lt;br /&gt;
cd $SLURM_SUBMIT_DIR&lt;br /&gt;
&lt;br /&gt;
module load NiaEnv/2019b&lt;br /&gt;
module load intel/2019u4&lt;br /&gt;
module load openmpi/4.0.1&lt;br /&gt;
&lt;br /&gt;
mpirun ./mpi_example&lt;br /&gt;
# or &amp;quot;srun ./mpi_example&amp;quot;&lt;br /&gt;
&amp;lt;/source&amp;gt;&lt;br /&gt;
Submit this script from your scratch directory with the command:&lt;br /&gt;
&lt;br /&gt;
    nia-login07:scratch$ sbatch mpi_job.sh&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;First line indicates that this is a bash script.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Lines starting with &amp;lt;code&amp;gt;#SBATCH&amp;lt;/code&amp;gt; go to SLURM.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;sbatch reads these lines as a job request (which it gives the name &amp;lt;code&amp;gt;mpi_job&amp;lt;/code&amp;gt;)&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;In this case, SLURM looks for 2 nodes each running 40 tasks (for a total of 80 tasks), for 1 hour&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Note that the mpifun flag &amp;quot;--ppn&amp;quot; (processors per node) is ignored.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Once it found such a node, it runs the script:&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Change to the submission directory;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Loads modules;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Runs the &amp;lt;code&amp;gt;mpi_example&amp;lt;/code&amp;gt; application (SLURM will inform mpirun or srun on how many processes to run).&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;To use hyperthreading, just change &amp;lt;code&amp;gt;--ntasks-per-node=40&amp;lt;/code&amp;gt; to &amp;lt;code&amp;gt;--ntasks-per-node=80&amp;lt;/code&amp;gt;, and add &amp;lt;code&amp;gt;--bind-to none&amp;lt;/code&amp;gt; to the mpirun command (the latter is necessary for OpenMPI only, not when using IntelMPI).&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Example submission script (OpenMP) ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;source lang=&amp;quot;bash&amp;quot;&amp;gt;#!/bin/bash&lt;br /&gt;
#SBATCH --nodes=1&lt;br /&gt;
#SBATCH --cpus-per-task=40&lt;br /&gt;
#SBATCH --time=1:00:00&lt;br /&gt;
#SBATCH --job-name=openmp_job&lt;br /&gt;
#SBATCH --output=openmp_output_%j.txt&lt;br /&gt;
#SBATCH --mail-type=FAIL&lt;br /&gt;
&lt;br /&gt;
cd $SLURM_SUBMIT_DIR&lt;br /&gt;
&lt;br /&gt;
module load NiaEnv/2019b&lt;br /&gt;
module load intel/2019u4&lt;br /&gt;
&lt;br /&gt;
export OMP_NUM_THREADS=$SLURM_CPUS_PER_TASK&lt;br /&gt;
&lt;br /&gt;
./openmp_example&lt;br /&gt;
# or &amp;quot;srun ./openmp_example&amp;quot;.&lt;br /&gt;
&amp;lt;/source&amp;gt;&lt;br /&gt;
Submit this script from your scratch directory with the command:&lt;br /&gt;
&lt;br /&gt;
    nia-login07:~$ sbatch openmp_job.sh&lt;br /&gt;
&lt;br /&gt;
* First line indicates that this is a bash script.&lt;br /&gt;
* Lines starting with &amp;lt;code&amp;gt;#SBATCH&amp;lt;/code&amp;gt; go to SLURM.&lt;br /&gt;
* sbatch reads these lines as a job request (which it gives the name &amp;lt;code&amp;gt;openmp_job&amp;lt;/code&amp;gt;) .&lt;br /&gt;
* In this case, SLURM looks for one node with 40 cores to be run inside one task, for 1 hour.&lt;br /&gt;
* Once it found such a node, it runs the script:&lt;br /&gt;
** Change to the submission directory;&lt;br /&gt;
** Loads modules;&lt;br /&gt;
** Sets an environment variable;&lt;br /&gt;
** Runs the &amp;lt;code&amp;gt;openmp_example&amp;lt;/code&amp;gt; application.&lt;br /&gt;
* To use hyperthreading, just change &amp;lt;code&amp;gt;--cpus-per-task=40&amp;lt;/code&amp;gt; to &amp;lt;code&amp;gt;--cpus-per-task=80&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
== Monitoring queued jobs ==&lt;br /&gt;
&lt;br /&gt;
Once the job is incorporated into the queue, there are some commands you can use to monitor its progress.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;lt;code&amp;gt;squeue&amp;lt;/code&amp;gt; or &amp;lt;code&amp;gt;sqc&amp;lt;/code&amp;gt; (a caching version of squeue) to show the job queue (&amp;lt;code&amp;gt;squeue -u $USER&amp;lt;/code&amp;gt; for just your jobs);&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;lt;code&amp;gt;squeue -j JOBID&amp;lt;/code&amp;gt; to get information on a specific job&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;(alternatively, &amp;lt;code&amp;gt;scontrol show job JOBID&amp;lt;/code&amp;gt;, which is more verbose).&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;lt;code&amp;gt;squeue --start -j JOBID&amp;lt;/code&amp;gt; to get an estimate for when a job will run; these tend not to be very accurate predictions.&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;lt;code&amp;gt;scancel -i JOBID&amp;lt;/code&amp;gt; to cancel the job.&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;lt;code&amp;gt;jobperf JOBID&amp;lt;/code&amp;gt; to get an instantaneous view of the cpu and memory usage of the nodes of the job while it is running.&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;lt;code&amp;gt;sacct&amp;lt;/code&amp;gt; to get information on your recent jobs.&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Further instructions for monitoring your jobs can be found on the [[Slurm#Monitoring_jobs | Slurm page]].  The [https://my.scinet.utoronto.ca my.SciNet] site is also a very useful tool for monitoring your current and past usage.&lt;br /&gt;
&lt;br /&gt;
= Visualization =&lt;br /&gt;
Information about how to use visualization tools on Niagara is available on [[Visualization]] page.&lt;br /&gt;
&lt;br /&gt;
= Support =&lt;br /&gt;
&lt;br /&gt;
* [mailto:support@scinet.utoronto.ca support@scinet.utoronto.ca]&lt;br /&gt;
* [mailto:niagara@tech.alliancecan.ca niagara@tech.alliancecan.ca]&lt;/div&gt;</summary>
		<author><name>Rzon</name></author>
	</entry>
	<entry>
		<id>https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7658</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7658"/>
		<updated>2026-04-14T14:51:26Z</updated>

		<summary type="html">&lt;p&gt;Rzon: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;__NOTOC__&lt;br /&gt;
{| style=&amp;quot;border-spacing:10px; width: 95%&amp;quot;&lt;br /&gt;
| style=&amp;quot;padding:1em; padding-top:.1em; border:2px solid #0645ad; background-color:#f6f6f6; border-radius:7px&amp;quot;|&lt;br /&gt;
&lt;br /&gt;
==System Status==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- Use &amp;quot;Up&amp;quot;, &amp;quot;Partial&amp;quot; or &amp;quot;Down&amp;quot;; these are templates. --&amp;gt;&lt;br /&gt;
{|style=&amp;quot;width:100%&amp;quot; &lt;br /&gt;
|{{Up3 | Trillium|https://docs.alliancecan.ca/wiki/Trillium_Quickstart}}&lt;br /&gt;
|{{Up3 | OnDemand|https://docs.alliancecan.ca/wiki/Trillium_Open_OnDemand_Quickstart}}&lt;br /&gt;
|{{Up | Globus |Globus}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Up | HPSS|HPSS}}&lt;br /&gt;
|{{Up | Balam|Balam}}&lt;br /&gt;
|{{Up | S4H | S4H}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Up | Teach|Teach}}&lt;br /&gt;
|{{Up3 | File system|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Storage}}&lt;br /&gt;
|{{Up3 | External Network|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Logging_in}} &lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Tue Apr 14, 2026, 10:30 am:''' The Trillium globus endpoint is working again.&lt;br /&gt;
&lt;br /&gt;
'''Sat Apr 11, 2026, 10:00 pm:''' The Trillium globus endpoint is not operational (it times out). We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''Thu Apr 09, 2026, 10:30 am:''' tri-dm4.scinet.utoronto.ca and robot4.scinet.utoronto.ca are in maintenance. Use 1,2, or 3 instead.&lt;br /&gt;
&lt;br /&gt;
'''Wed Apr 08, 2026, 6:30 pm:''' Software from the CVMFS 'restricted' area is back.&lt;br /&gt;
&lt;br /&gt;
'''Wed Apr 08, 2026, 1:40 pm:''' Software from the CVMFS 'restricted' area is not available on many of the Trillium and Open OnDemand nodes. We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''Mon Apr 06, 2026, 10:00 pm:''' We will have to reschedule the HPSS update. This attempt didn't work as expected&lt;br /&gt;
&lt;br /&gt;
'''Mon Apr 06, 2026, 8:00 pm:''' HPSS scheduled maintenance: update of HPSS to v11.3_u4 and hsi-htar to v11.3_u1 (bug fixes)&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--  When removing system status entries, please archive them to: --&amp;gt;&lt;br /&gt;
[[Previous messages]]&lt;br /&gt;
&lt;br /&gt;
{|style=&amp;quot;border-spacing: 10px;width: 100%&amp;quot;&lt;br /&gt;
|valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== QuickStart Guides ==&lt;br /&gt;
* [https://docs.alliancecan.ca/wiki/Trillium_Quickstart Trillium Quickstart]&lt;br /&gt;
* [[HPSS | HPSS archival storage]]&lt;br /&gt;
* [[Teach|Teach cluster]]&lt;br /&gt;
* [[FAQ | FAQ (frequently asked questions)]]&lt;br /&gt;
* [[Acknowledging SciNet]]&lt;br /&gt;
| valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== Tutorials, Manuals, etc. ==&lt;br /&gt;
* [https://education.scinet.utoronto.ca SciNet education material]&lt;br /&gt;
* [https://www.youtube.com/c/SciNetHPCattheUniversityofToronto SciNet's YouTube channel]&lt;br /&gt;
* [[Modules for Mist]] &lt;br /&gt;
* [[Commercial software]]&lt;br /&gt;
* [[Burst Buffer]]&lt;br /&gt;
* [[SSH#SSH Keys|SSH keys]]&lt;br /&gt;
* [[SSH Tunneling]]&lt;br /&gt;
* [[Visualization]]&lt;br /&gt;
* [[Running Serial Jobs on Niagara]]&lt;br /&gt;
* [[Jupyter Hub]]&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Rzon</name></author>
	</entry>
	<entry>
		<id>https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7655</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7655"/>
		<updated>2026-04-13T18:17:08Z</updated>

		<summary type="html">&lt;p&gt;Rzon: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;__NOTOC__&lt;br /&gt;
{| style=&amp;quot;border-spacing:10px; width: 95%&amp;quot;&lt;br /&gt;
| style=&amp;quot;padding:1em; padding-top:.1em; border:2px solid #0645ad; background-color:#f6f6f6; border-radius:7px&amp;quot;|&lt;br /&gt;
&lt;br /&gt;
==System Status==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- Use &amp;quot;Up&amp;quot;, &amp;quot;Partial&amp;quot; or &amp;quot;Down&amp;quot;; these are templates. --&amp;gt;&lt;br /&gt;
{|style=&amp;quot;width:100%&amp;quot; &lt;br /&gt;
|{{Up3 | Trillium|https://docs.alliancecan.ca/wiki/Trillium_Quickstart}}&lt;br /&gt;
|{{Up3 | OnDemand|https://docs.alliancecan.ca/wiki/Trillium_Open_OnDemand_Quickstart}}&lt;br /&gt;
|{{Partial | Globus |Globus}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Up | HPSS|HPSS}}&lt;br /&gt;
|{{Up | Balam|Balam}}&lt;br /&gt;
|{{Up | S4H | S4H}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Up | Teach|Teach}}&lt;br /&gt;
|{{Up3 | File system|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Storage}}&lt;br /&gt;
|{{Up3 | External Network|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Logging_in}} &lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Sat Apr 11, 2026, 10:00 pm:''' The Trillium globus endpoint is not operational (it times out). We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''Thu Apr 09, 2026, 10:30 am:''' tri-dm4.scinet.utoronto.ca and robot4.scinet.utoronto.ca are in maintenance. Use 1,2, or 3 instead.&lt;br /&gt;
&lt;br /&gt;
'''Wed Apr 08, 2026, 6:30 pm:''' Software from the CVMFS 'restricted' area is back.&lt;br /&gt;
&lt;br /&gt;
'''Wed Apr 08, 2026, 1:40 pm:''' Software from the CVMFS 'restricted' area is not available on many of the Trillium and Open OnDemand nodes. We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''Mon Apr 06, 2026, 10:00 pm:''' We will have to reschedule the HPSS update. This attempt didn't work as expected&lt;br /&gt;
&lt;br /&gt;
'''Mon Apr 06, 2026, 8:00 pm:''' HPSS scheduled maintenance: update of HPSS to v11.3_u4 and hsi-htar to v11.3_u1 (bug fixes)&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--  When removing system status entries, please archive them to: --&amp;gt;&lt;br /&gt;
[[Previous messages]]&lt;br /&gt;
&lt;br /&gt;
{|style=&amp;quot;border-spacing: 10px;width: 100%&amp;quot;&lt;br /&gt;
|valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== QuickStart Guides ==&lt;br /&gt;
* [https://docs.alliancecan.ca/wiki/Trillium_Quickstart Trillium Quickstart]&lt;br /&gt;
* [[HPSS | HPSS archival storage]]&lt;br /&gt;
* [[Teach|Teach cluster]]&lt;br /&gt;
* [[FAQ | FAQ (frequently asked questions)]]&lt;br /&gt;
* [[Acknowledging SciNet]]&lt;br /&gt;
| valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== Tutorials, Manuals, etc. ==&lt;br /&gt;
* [https://education.scinet.utoronto.ca SciNet education material]&lt;br /&gt;
* [https://www.youtube.com/c/SciNetHPCattheUniversityofToronto SciNet's YouTube channel]&lt;br /&gt;
* [[Modules for Mist]] &lt;br /&gt;
* [[Commercial software]]&lt;br /&gt;
* [[Burst Buffer]]&lt;br /&gt;
* [[SSH#SSH Keys|SSH keys]]&lt;br /&gt;
* [[SSH Tunneling]]&lt;br /&gt;
* [[Visualization]]&lt;br /&gt;
* [[Running Serial Jobs on Niagara]]&lt;br /&gt;
* [[Jupyter Hub]]&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Rzon</name></author>
	</entry>
	<entry>
		<id>https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7652</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7652"/>
		<updated>2026-04-09T20:30:31Z</updated>

		<summary type="html">&lt;p&gt;Rzon: /* System Status */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;__NOTOC__&lt;br /&gt;
{| style=&amp;quot;border-spacing:10px; width: 95%&amp;quot;&lt;br /&gt;
| style=&amp;quot;padding:1em; padding-top:.1em; border:2px solid #0645ad; background-color:#f6f6f6; border-radius:7px&amp;quot;|&lt;br /&gt;
&lt;br /&gt;
==System Status==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- Use &amp;quot;Up&amp;quot;, &amp;quot;Partial&amp;quot; or &amp;quot;Down&amp;quot;; these are templates. --&amp;gt;&lt;br /&gt;
{|style=&amp;quot;width:100%&amp;quot; &lt;br /&gt;
|{{Up3 | Trillium|https://docs.alliancecan.ca/wiki/Trillium_Quickstart}}&lt;br /&gt;
|{{Up3 | OnDemand|https://docs.alliancecan.ca/wiki/Trillium_Open_OnDemand_Quickstart}}&lt;br /&gt;
|{{Up | Globus |Globus}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Up | HPSS|HPSS}}&lt;br /&gt;
|{{Up | Balam|Balam}}&lt;br /&gt;
|{{Up | S4H | S4H}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Up | Teach|Teach}}&lt;br /&gt;
|{{Up3 | File system|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Storage}}&lt;br /&gt;
|{{Up3 | External Network|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Logging_in}} &lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Thu Apr 09, 2026, 10:30 am:''' tri-dm4.scinet.utoronto.ca and robot4.scinet.utoronto.ca are in maintenance. Use 1,2, or 3 instead.&lt;br /&gt;
&lt;br /&gt;
'''Wed Apr 08, 2026, 6:30 pm:''' Software from the CVMFS 'restricted' area is back.&lt;br /&gt;
&lt;br /&gt;
'''Wed Apr 08, 2026, 1:40 pm:''' Software from the CVMFS 'restricted' area is not available on many of the Trillium and Open OnDemand nodes. We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''Mon Apr 06, 2026, 10:00 pm:''' We will have to reschedule the HPSS update. This attempt didn't work as expected&lt;br /&gt;
&lt;br /&gt;
'''Mon Apr 06, 2026, 8:00 pm:''' HPSS scheduled maintenance: update of HPSS to v11.3_u4 and hsi-htar to v11.3_u1 (bug fixes)&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--  When removing system status entries, please archive them to: --&amp;gt;&lt;br /&gt;
[[Previous messages]]&lt;br /&gt;
&lt;br /&gt;
{|style=&amp;quot;border-spacing: 10px;width: 100%&amp;quot;&lt;br /&gt;
|valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== QuickStart Guides ==&lt;br /&gt;
* [https://docs.alliancecan.ca/wiki/Trillium_Quickstart Trillium Quickstart]&lt;br /&gt;
* [[HPSS | HPSS archival storage]]&lt;br /&gt;
* [[Teach|Teach cluster]]&lt;br /&gt;
* [[FAQ | FAQ (frequently asked questions)]]&lt;br /&gt;
* [[Acknowledging SciNet]]&lt;br /&gt;
| valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== Tutorials, Manuals, etc. ==&lt;br /&gt;
* [https://education.scinet.utoronto.ca SciNet education material]&lt;br /&gt;
* [https://www.youtube.com/c/SciNetHPCattheUniversityofToronto SciNet's YouTube channel]&lt;br /&gt;
* [[Modules for Mist]] &lt;br /&gt;
* [[Commercial software]]&lt;br /&gt;
* [[Burst Buffer]]&lt;br /&gt;
* [[SSH#SSH Keys|SSH keys]]&lt;br /&gt;
* [[SSH Tunneling]]&lt;br /&gt;
* [[Visualization]]&lt;br /&gt;
* [[Running Serial Jobs on Niagara]]&lt;br /&gt;
* [[Jupyter Hub]]&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Rzon</name></author>
	</entry>
	<entry>
		<id>https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7649</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7649"/>
		<updated>2026-04-08T23:16:28Z</updated>

		<summary type="html">&lt;p&gt;Rzon: /* System Status */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;__NOTOC__&lt;br /&gt;
{| style=&amp;quot;border-spacing:10px; width: 95%&amp;quot;&lt;br /&gt;
| style=&amp;quot;padding:1em; padding-top:.1em; border:2px solid #0645ad; background-color:#f6f6f6; border-radius:7px&amp;quot;|&lt;br /&gt;
&lt;br /&gt;
==System Status==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- Use &amp;quot;Up&amp;quot;, &amp;quot;Partial&amp;quot; or &amp;quot;Down&amp;quot;; these are templates. --&amp;gt;&lt;br /&gt;
{|style=&amp;quot;width:100%&amp;quot; &lt;br /&gt;
|{{Up3 | Trillium|https://docs.alliancecan.ca/wiki/Trillium_Quickstart}}&lt;br /&gt;
|{{Up3 | OnDemand|https://docs.alliancecan.ca/wiki/Trillium_Open_OnDemand_Quickstart}}&lt;br /&gt;
|{{Up | Globus |Globus}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Up | HPSS|HPSS}}&lt;br /&gt;
|{{Up | Balam|Balam}}&lt;br /&gt;
|{{Up | S4H | S4H}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Up | Teach|Teach}}&lt;br /&gt;
|{{Up3 | File system|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Storage}}&lt;br /&gt;
|{{Up3 | External Network|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Logging_in}} &lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Wed Apr 08, 2026, 6:30 pm:''' Software from the CVMFS 'restricted' area is back.&lt;br /&gt;
&lt;br /&gt;
'''Wed Apr 08, 2026, 1:40 pm:''' Software from the CVMFS 'restricted' area is not available on many of the Trillium and Open OnDemand nodes. We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''Mon Apr 06, 2026, 10:00 pm:''' We will have to reschedule the HPSS update. This attempt didn't work as expected&lt;br /&gt;
&lt;br /&gt;
'''Mon Apr 06, 2026, 8:00 pm:''' HPSS scheduled maintenance: update of HPSS to v11.3_u4 and hsi-htar to v11.3_u1 (bug fixes)&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--  When removing system status entries, please archive them to: --&amp;gt;&lt;br /&gt;
[[Previous messages]]&lt;br /&gt;
&lt;br /&gt;
{|style=&amp;quot;border-spacing: 10px;width: 100%&amp;quot;&lt;br /&gt;
|valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== QuickStart Guides ==&lt;br /&gt;
* [https://docs.alliancecan.ca/wiki/Trillium_Quickstart Trillium Quickstart]&lt;br /&gt;
* [[HPSS | HPSS archival storage]]&lt;br /&gt;
* [[Teach|Teach cluster]]&lt;br /&gt;
* [[FAQ | FAQ (frequently asked questions)]]&lt;br /&gt;
* [[Acknowledging SciNet]]&lt;br /&gt;
| valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== Tutorials, Manuals, etc. ==&lt;br /&gt;
* [https://education.scinet.utoronto.ca SciNet education material]&lt;br /&gt;
* [https://www.youtube.com/c/SciNetHPCattheUniversityofToronto SciNet's YouTube channel]&lt;br /&gt;
* [[Modules for Mist]] &lt;br /&gt;
* [[Commercial software]]&lt;br /&gt;
* [[Burst Buffer]]&lt;br /&gt;
* [[SSH#SSH Keys|SSH keys]]&lt;br /&gt;
* [[SSH Tunneling]]&lt;br /&gt;
* [[Visualization]]&lt;br /&gt;
* [[Running Serial Jobs on Niagara]]&lt;br /&gt;
* [[Jupyter Hub]]&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Rzon</name></author>
	</entry>
	<entry>
		<id>https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7646</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7646"/>
		<updated>2026-04-08T20:49:34Z</updated>

		<summary type="html">&lt;p&gt;Rzon: /* System Status */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;__NOTOC__&lt;br /&gt;
{| style=&amp;quot;border-spacing:10px; width: 95%&amp;quot;&lt;br /&gt;
| style=&amp;quot;padding:1em; padding-top:.1em; border:2px solid #0645ad; background-color:#f6f6f6; border-radius:7px&amp;quot;|&lt;br /&gt;
&lt;br /&gt;
==System Status==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- Use &amp;quot;Up&amp;quot;, &amp;quot;Partial&amp;quot; or &amp;quot;Down&amp;quot;; these are templates. --&amp;gt;&lt;br /&gt;
{|style=&amp;quot;width:100%&amp;quot; &lt;br /&gt;
|{{Up3 | Trillium|https://docs.alliancecan.ca/wiki/Trillium_Quickstart}}&lt;br /&gt;
|{{Up3 | OnDemand|https://docs.alliancecan.ca/wiki/Trillium_Open_OnDemand_Quickstart}}&lt;br /&gt;
|{{Up | Globus |Globus}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Up | HPSS|HPSS}}&lt;br /&gt;
|{{Up | Balam|Balam}}&lt;br /&gt;
|{{Up | S4H | S4H}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Up | Teach|Teach}}&lt;br /&gt;
|{{Up3 | File system|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Storage}}&lt;br /&gt;
|{{Up3 | External Network|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Logging_in}} &lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Wed Apr 08, 2026, 1:40 pm:''' Software from the CVMFS 'restricted' area is not available on many of the Trillium and Open OnDemand nodes. We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''Mon Apr 06, 2026, 10:00 pm:''' We will have to reschedule the HPSS update. This attempt didn't work as expected&lt;br /&gt;
&lt;br /&gt;
'''Mon Apr 06, 2026, 8:00 pm:''' HPSS scheduled maintenance: update of HPSS to v11.3_u4 and hsi-htar to v11.3_u1 (bug fixes)&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--  When removing system status entries, please archive them to: --&amp;gt;&lt;br /&gt;
[[Previous messages]]&lt;br /&gt;
&lt;br /&gt;
{|style=&amp;quot;border-spacing: 10px;width: 100%&amp;quot;&lt;br /&gt;
|valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== QuickStart Guides ==&lt;br /&gt;
* [https://docs.alliancecan.ca/wiki/Trillium_Quickstart Trillium Quickstart]&lt;br /&gt;
* [[HPSS | HPSS archival storage]]&lt;br /&gt;
* [[Teach|Teach cluster]]&lt;br /&gt;
* [[FAQ | FAQ (frequently asked questions)]]&lt;br /&gt;
* [[Acknowledging SciNet]]&lt;br /&gt;
| valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== Tutorials, Manuals, etc. ==&lt;br /&gt;
* [https://education.scinet.utoronto.ca SciNet education material]&lt;br /&gt;
* [https://www.youtube.com/c/SciNetHPCattheUniversityofToronto SciNet's YouTube channel]&lt;br /&gt;
* [[Modules for Mist]] &lt;br /&gt;
* [[Commercial software]]&lt;br /&gt;
* [[Burst Buffer]]&lt;br /&gt;
* [[SSH#SSH Keys|SSH keys]]&lt;br /&gt;
* [[SSH Tunneling]]&lt;br /&gt;
* [[Visualization]]&lt;br /&gt;
* [[Running Serial Jobs on Niagara]]&lt;br /&gt;
* [[Jupyter Hub]]&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Rzon</name></author>
	</entry>
	<entry>
		<id>https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7634</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7634"/>
		<updated>2026-04-02T17:57:10Z</updated>

		<summary type="html">&lt;p&gt;Rzon: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;__NOTOC__&lt;br /&gt;
{| style=&amp;quot;border-spacing:10px; width: 95%&amp;quot;&lt;br /&gt;
| style=&amp;quot;padding:1em; padding-top:.1em; border:2px solid #0645ad; background-color:#f6f6f6; border-radius:7px&amp;quot;|&lt;br /&gt;
&lt;br /&gt;
==System Status==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- Use &amp;quot;Up&amp;quot;, &amp;quot;Partial&amp;quot; or &amp;quot;Down&amp;quot;; these are templates. --&amp;gt;&lt;br /&gt;
{|style=&amp;quot;width:100%&amp;quot; &lt;br /&gt;
|{{Up3 | Trillium|https://docs.alliancecan.ca/wiki/Trillium_Quickstart}}&lt;br /&gt;
|{{Up3 | OnDemand|https://docs.alliancecan.ca/wiki/Trillium_Open_OnDemand_Quickstart}}&lt;br /&gt;
|{{Up | Globus |Globus}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Up | HPSS|HPSS}}&lt;br /&gt;
|{{Up | Balam|Balam}}&lt;br /&gt;
|{{Up | S4H | S4H}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Up | Teach|Teach}}&lt;br /&gt;
|{{Up3 | File system|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Storage}}&lt;br /&gt;
|{{Up3 | External Network|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Logging_in}} &lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Mon Apr 06, 2026, 8:00 pm:''' HPSS scheduled maintenance: update of HPSS to v11.3_u4 and hsi-htar to v11.3_u1 (bug fixes)&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--  When removing system status entries, please archive them to: --&amp;gt;&lt;br /&gt;
[[Previous messages]]&lt;br /&gt;
&lt;br /&gt;
{|style=&amp;quot;border-spacing: 10px;width: 100%&amp;quot;&lt;br /&gt;
|valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== QuickStart Guides ==&lt;br /&gt;
* [https://docs.alliancecan.ca/wiki/Trillium_Quickstart Trillium Quickstart]&lt;br /&gt;
* [[HPSS | HPSS archival storage]]&lt;br /&gt;
* [[Teach|Teach cluster]]&lt;br /&gt;
* [[FAQ | FAQ (frequently asked questions)]]&lt;br /&gt;
* [[Acknowledging SciNet]]&lt;br /&gt;
| valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== Tutorials, Manuals, etc. ==&lt;br /&gt;
* [https://education.scinet.utoronto.ca SciNet education material]&lt;br /&gt;
* [https://www.youtube.com/c/SciNetHPCattheUniversityofToronto SciNet's YouTube channel]&lt;br /&gt;
* [[Modules for Mist]] &lt;br /&gt;
* [[Commercial software]]&lt;br /&gt;
* [[Burst Buffer]]&lt;br /&gt;
* [[SSH#SSH Keys|SSH keys]]&lt;br /&gt;
* [[SSH Tunneling]]&lt;br /&gt;
* [[Visualization]]&lt;br /&gt;
* [[Running Serial Jobs on Niagara]]&lt;br /&gt;
* [[Jupyter Hub]]&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Rzon</name></author>
	</entry>
	<entry>
		<id>https://docs.scinet.utoronto.ca/index.php?title=Previous_messages&amp;diff=7631</id>
		<title>Previous messages</title>
		<link rel="alternate" type="text/html" href="https://docs.scinet.utoronto.ca/index.php?title=Previous_messages&amp;diff=7631"/>
		<updated>2026-04-02T17:56:57Z</updated>

		<summary type="html">&lt;p&gt;Rzon: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;'''Wed Mar 25, 2026, 5:00 pm:''' Trillium is operational again.&lt;br /&gt;
&lt;br /&gt;
'''Wed Mar 25, 2026, 9:00 am:''' Teach is operational again.&lt;br /&gt;
&lt;br /&gt;
'''Tue Mar 24, 2026, 8:45 pm:''' Open OnDemand is operational again.&lt;br /&gt;
&lt;br /&gt;
'''Tue Mar 24, 2026, 1:00 pm:''' External connectivity is back. &lt;br /&gt;
&lt;br /&gt;
'''Tue Mar 24, 2026, 12:05 pm:''' External connectivity to the data centre was lost. &lt;br /&gt;
&lt;br /&gt;
'''Tue Mar 24, 2026, 7:00 am:''' Maintenance has started.&lt;br /&gt;
&lt;br /&gt;
'''Mon Mar 16, 2026. 13:30pm''' Recovering.  Almost all systems are up again. Please resubmit your jobs that crashed. &lt;br /&gt;
&lt;br /&gt;
'''Mon Mar 16, 2026. 12:00pm''' Power glitch at the data centre caused compute nodes to go down.&lt;br /&gt;
&lt;br /&gt;
'''Thu Mar 12, 2026, 4:15 pm''' Connection to Trillium are operational again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Mar 12, 2026, 1:00 pm''' We've had some login issues particularly for Trillium-GPU. We're investigating.&lt;br /&gt;
&lt;br /&gt;
'''Downtime Announcement:'''  The winter cooling tower maintenance for the SciNet data centre will take place on March 24 and 25, 2026, starting at 7:00 a.m. on the 24th.  All SciNet systems (Trillium, OnDemand, Balam, S4H, Teach, as well as hosted equipment) will have their compute nodes shut down. Login nodes, file systems, and the HPSS system will remain available, and&lt;br /&gt;
jobs will be held in the queue until maintenance is complete.  Starting 7am on Mar 23, users are encouraged to submit small and short jobs that may be scheduled before the maintenance begins.&lt;br /&gt;
&lt;br /&gt;
'''Fri Feb 20, 2026, 11:35 pm:''' Power glitch, ~480 compute nodes rebooted. Regional power quality has been quite poor lately ([https://www.yorkregion.com/news/road-salt-blamed-for-power-outages/article_1a36d25d-5f97-56ee-a0c7-c49c7b732d38.html 1],&lt;br /&gt;
[https://www.yorkregion.com/news/power-company-executive-responds-to-york-region-outages/article_c4d072e7-2892-5c9c-8deb-ac5e1936779c.html 2]).&lt;br /&gt;
&lt;br /&gt;
'''Thu Feb 19, 2026, 3:00 pm:''' Systems restored. Please report issues to support@scinet.utoronto.ca.&lt;br /&gt;
&lt;br /&gt;
'''Tue Feb 17, 2026, 8:40 am:''' Power outage at the data centre.  Cooling issues have developed as a result.  Major systems (Trillium, S4H) are expected to be down until sometime Thursday. Login nodes and file systems will remain accessible.&lt;br /&gt;
&lt;br /&gt;
'''Mon Feb 16, 2026, 8:40 pm:''' Electricity is unstable in the data centre area due to severe snowfall.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 1:40 pm:''' All services are operational again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 12:00 pm:''' The Trillium and Open OnDemand compute nodes are operational again. We are still working on bringing Balam, Neptune and S4H nodes up again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 10:00 am:''' There was a power glitch at the data centre overnight. The login nodes are accessible but the compute nodes are down.  &lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 16, 2026, 11:00 pm:''' HPSS is back online, and accessible via alliancecan#hpss Globus endpoint. &lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 15, 2026, 10:00 pm:''' HPSS will undergo maintenance on Friday morning, Jan/16/2025, , including alliancecan#hpss Globus endpoint &lt;br /&gt;
&lt;br /&gt;
'''Tue Jan 6, 2026, 10:15 am:''' OnDemand has been fixed and is working again.&lt;br /&gt;
&lt;br /&gt;
'''Mon Jan 5, 2026, 9:00 pm:''' The authentication mechanism of OnDemand is not working.&lt;br /&gt;
&lt;br /&gt;
'''Wed Dec 31, 2025, 12:40 pm:''' We believe the problem has now been resolved.  Please let us know if you still experience login problems or aborted jobs.&lt;br /&gt;
&lt;br /&gt;
'''Tue Dec 30, 2025, 2:10 pm:''' We are experiencing problems with authentication, resulting in failed logins, OOD errors, and aborted jobs (with &amp;quot;prolog error&amp;quot;).  Please bear with us, as we are very short-staffed during the holiday break.  We will post updates here.&lt;br /&gt;
&lt;br /&gt;
'''Tue Dec 3, 2025, 11:30 am:''' Open OnDemand is fully operational again.&lt;br /&gt;
&lt;br /&gt;
'''Sat Nov 29, 2025, 00:40 am:''' There has been a problem with the water chiller. Some systems are offline.&lt;br /&gt;
&lt;br /&gt;
'''Wed Nov 5, 2025, 12:55 pm:''' Balam is back online.&lt;br /&gt;
&lt;br /&gt;
'''Wed Nov 5, 2025, 10:00 am:''' Open OnDemand is back online.&lt;br /&gt;
&lt;br /&gt;
'''Tue Nov 4, 2025, 11:00 pm:''' Most of the work is done, data movers, Globus, and HPSS are back online. Remaining services will be worked on tomorrow.&lt;br /&gt;
&lt;br /&gt;
'''Tue Nov 4, 2025, 8:30 am:''' Scheduled network maintenance. Trillium cluster is *not* affected.&lt;br /&gt;
&lt;br /&gt;
'''Tue Oct 21, 2025, 17:30 am:''' Balam maintenance finished.&lt;br /&gt;
&lt;br /&gt;
'''Tue Oct 21, 2025, 7:00 am:''' Balam maintenance day.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 15, 2025, 3:55 pm:''' Trillium inbound connections through trillium.alliancecan.ca or trillium.scinet.utoronto.ca are working again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 15, 2025, 3:05 pm:''' Trillium is experiencing external network issues for both incoming traffic. Please try: ssh USERNAME@tri-login01.scinet.utoronto.ca in the meantime.&lt;br /&gt;
 &lt;br /&gt;
'''Thu Oct 06, 2025, 8:00 pm:''' HPSS is fully functional. You may submit archive jobs from trillium login nodes, datamovers and robots.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 03, 2025, 6:30 pm:''' HPSS is back online, and already accessible via alliancecan#hpss Globus endpoint. Directory tree now follows the other Alliance clusters. We're still working on job submission via Slurm&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 01, 2025, 0:00 am:''' Niagara compute nodes are now unavailable for regular users. The login nodes will remain available for a while to allow a few last data transfers, although transfers from the Niagara file systems to Trillium are best done on nia-dm1.scinet.utoronto.ca.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 01, 2025, 9:30 am:''' HPSS is down for scheduled maintenance, including alliancecan#hpss Globus endpoint&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
'''Thu Sep 18, 2025, 11:30 am:''' Open OnDemand is fully functional again.&lt;br /&gt;
&lt;br /&gt;
'''Wed Sep 17, 2025, 6:00 pm:''' Niagara is back up as well (including its Globus endpoint).  We are still working on the other systems.&lt;br /&gt;
&lt;br /&gt;
'''Wed Sep 17, 2025, 1:40 pm:''' Trillium is back up (except for its Globus endpoint).  We are working on the other systems still.&lt;br /&gt;
&lt;br /&gt;
'''Tue Sep 16, 2025, 5:45 pm:''' Unfortunately, we cannot bring all systems up yet because we are waiting for a spare part for the cooling system that will be brought tomorrow.  In the meantime, we have managed to keep the Trillium login nodes up, but not other systems.&lt;br /&gt;
&lt;br /&gt;
'''Tue Sep 16, 2025, from 7:00 am to 5:00 pm (EDT):''' The SciNet datacentre will undergo maintenance of several critical parts of the centre.  This will require a full shutdown of all SciNet systems (Trillium, Niagara, Mist, HPSS, Rouge, Teach, as well as hosted equipment). This will also be the time that the Mist cluster gets decommissioned. &lt;br /&gt;
&lt;br /&gt;
'''Fri Sep 12 22:03:17 EDT 2025:''' HPSS software and OS upgrades are finished.&lt;br /&gt;
&lt;br /&gt;
'''Tue Sep  9 17:05:38 EDT 2025:''' Starting tomorrow, Sep/10, and for the following 3 days HPSS will be down for software and OS upgrades. We will strive to finish sooner, at which time we will make the system available to users again.&lt;br /&gt;
&lt;br /&gt;
===Mist/Niagara Decommissioning Schedule===&lt;br /&gt;
&lt;br /&gt;
'''September 4, 2025'''&lt;br /&gt;
* Niagara reduced to 863 compute nodes.&lt;br /&gt;
&lt;br /&gt;
'''September 9, 2025'''&lt;br /&gt;
* Niagara's Open OnDemand decommissioned.&lt;br /&gt;
* Brief data centre connection outage at 9 AM EDT&lt;br /&gt;
* Niagara reduced to 647 compute nodes at end of day.&lt;br /&gt;
&lt;br /&gt;
'''September 11, 2025'''&lt;br /&gt;
* Trillium Open OnDemand goes live.&lt;br /&gt;
&lt;br /&gt;
'''September 16, 2025'''&lt;br /&gt;
* '''Full-day data centre maintenance'''&lt;br /&gt;
* Niagara reduced to 431 compute nodes.&lt;br /&gt;
* Mist decommissioned.&lt;br /&gt;
&lt;br /&gt;
'''September 24, 2025'''&lt;br /&gt;
* Niagara reduced to 215 compute nodes.&lt;br /&gt;
&lt;br /&gt;
'''September 30, 2025'''&lt;br /&gt;
* Niagara decommissioned.&lt;br /&gt;
&lt;br /&gt;
'''August 25, 2025, 9:50 EDT:''' Open Ondemand is fully operational again.&lt;br /&gt;
&lt;br /&gt;
'''August 22, 2025, 3:15 PM EDT:''' Open Ondemand has issues launching new interactive apps. We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''August 20, 2025, 10:00 AM EDT:''' The GPU scheduler on Trillium is fully operational again.&lt;br /&gt;
&lt;br /&gt;
'''August 19, 2025, 5:00 PM EDT:''' The GPU scheduler on Trillium has trouble scheduling multi-GPU jobs.  We're investigating the issue.&lt;br /&gt;
&lt;br /&gt;
'''August 7, 2025:''' CVMFS issues are resolved.&lt;br /&gt;
&lt;br /&gt;
'''August 6, 2025:''' We are seeing intermittent issues with the software on CVMFS on Niagara. We're investigating.&lt;br /&gt;
&lt;br /&gt;
'''July 31, 2025, 4:00 PM EDT - 5:00 PM EDT:''' As announced, all systems connected to the Niagara file system (Mist, Niagara, HPSS, Balam, and Rouge) will be paused and inaccessible for one hour to start the transfer of files from the Niagara file system to the Trillium file system. &lt;br /&gt;
&lt;br /&gt;
'''January 6, 2025:''' As part of the installation of the new computing cluster Trillium, there is now a permanent reduction in computing capacity of Niagara to 50% and of Mist to 35%.&lt;br /&gt;
&lt;br /&gt;
'''July 9, 2025:''' The [[Teach]] cluster will be unavailable for the day for network maintenance.&lt;br /&gt;
&lt;br /&gt;
'''July 4, 2025:''' Open OnDemand is back up.&lt;br /&gt;
&lt;br /&gt;
'''July 4, 2025:''' Open OnDemand is down. We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''June 25, 2025, 7:15 PM EDT:''' The [[Teach]] cluster's scheduler is up again.&lt;br /&gt;
&lt;br /&gt;
'''June 25, 2025, 4:30 PM EDT:''' The [[Teach]] cluster's scheduler is down. We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''April 30, 2025, 9:30 AM EDT:''' The [[Teach]] cluster is available again.&lt;br /&gt;
&lt;br /&gt;
'''April 30, 2025:''' The [[Teach]] cluster will be unavailable from 8:00 am to about 12:00 noon for file system maintenance.&lt;br /&gt;
&lt;br /&gt;
'''April 1, 2025:''' The Jupyter Hub has been replaced by SciNet's [[Open OnDemand Quickstart|Open OnDemand service]].&lt;br /&gt;
&lt;br /&gt;
'''March 1, 2025:''' As of March 1st scratch purging is suspended until after Trillium comes online.&lt;br /&gt;
&lt;br /&gt;
'''April 15, 2025 12:40 pm EDT: '''Balam login node is accessible again.&lt;br /&gt;
&lt;br /&gt;
'''April 15, 2025 12:10 pm EDT: '''Balam login node is under maintenance and temporarily inaccessible to users.&lt;br /&gt;
&lt;br /&gt;
'''April 9, 2025 9PM:''' HPSS is back online.&lt;br /&gt;
&lt;br /&gt;
'''April 8, 2025 9PM:''' HPSS is being reserved for OS updates on April 9 (Wednesday).&lt;br /&gt;
&lt;br /&gt;
'''March 31, 2025 3:20 pm EDT:''' Mist login node is accessible again.&lt;br /&gt;
&lt;br /&gt;
'''March 31, 2025 2:45 pm EDT:''' Mist login node is under maintenance and temporarily inaccessible to users.&lt;br /&gt;
&lt;br /&gt;
'''March 28, 2025 3:00 pm - 4:00 pm EDT:''' A short maintenance was needed for the Teach compute nodes; you might have experienced some job scheduling delays on that cluster. &lt;br /&gt;
&lt;br /&gt;
'''March 20, 2025 10:30 am EDT:''' Teach compute nodes are back. &lt;br /&gt;
&lt;br /&gt;
'''March 19, 2025 11:00 pm EDT:''' Teach compute nodes are down again. &lt;br /&gt;
&lt;br /&gt;
'''March 19, 2025 5:15pm EDT:''' Maintenance of the cooling system was performed successfully. The cluster is back on line&lt;br /&gt;
&lt;br /&gt;
'''March 19, 2025 8:00 am - 5:00 pm EDT:''' Maintenance of the cooling system as well as preparations for the Trillium cluster will require a shutdown of the compute nodes of all SciNet systems (Niagara, Mist, Rouge, Balam, Teach, as well as hosted equipment). The login nodes, file systems and the HPSS system will remain available. The scheduler will hold jobs that are submitted until the maintenance has finished.&lt;br /&gt;
&lt;br /&gt;
'''March 18, 2025 10:00 am EDT:''' Teach compute nodes are back.&lt;br /&gt;
&lt;br /&gt;
'''March 17, 2025 10:00 pm EDT:''' Teach compute nodes are down. We are working on it. &lt;br /&gt;
&lt;br /&gt;
'''February 27, 2025 9:00 pm EST:''' Access to HPSS via Globus has been restored.&lt;br /&gt;
&lt;br /&gt;
'''February 25, 2025 2:30 pm EST:''' Access to HPSS via Globus is currently suspended (sorry, trivial upgrade has gone wrong).&lt;br /&gt;
&lt;br /&gt;
'''February 25, 2025 12:30 pm EST:''' Mist login node is accessible again.&lt;br /&gt;
&lt;br /&gt;
'''February 25, 2025 11:50 am EST:''' Mist login node is under maintenance and temporarily inaccessible to users.&lt;br /&gt;
&lt;br /&gt;
'''February 7, 2025 2:45 pm EST:''' Systems are back online.&lt;br /&gt;
&lt;br /&gt;
'''Fri Feb  7 01:04:33 EST 2025:''' There has been a problem with the water chiller. Automatic thermal shutdown of the compute nodes&lt;br /&gt;
&lt;br /&gt;
'''January 31, 2025 11:45 am EST:''' Power is back.&lt;br /&gt;
&lt;br /&gt;
'''January 31, 2025 6:00 am EST:''' Power outage in the data center. Many compute jobs will have stopped. Until power gets restored, parts of the systems are  running on the generator. No ETA on full power restoration.&lt;br /&gt;
 &lt;br /&gt;
'''January 28, 2025 9:30 pm EST:''' The CCEnv stack has been restored.&lt;br /&gt;
&lt;br /&gt;
'''January 28, 2025 5:00 pm EST:''' The CCEnv stack from cvmfs has issues and may not work reliable.&lt;br /&gt;
&lt;br /&gt;
'''January 23, 2025 9:00 am - 1:00 pm EST:''' Balam, Rouge and Neptune compute nodes will be shut down from 9 AM to 1 PM EST for additional electrical work.&lt;br /&gt;
&lt;br /&gt;
'''January 22, 2025 12:55 pm EST:''' Compute nodes are back online&lt;br /&gt;
&lt;br /&gt;
'''January 22, 2025 8:00 am - 5:00 pm EST:''' Preparations for the new system Trillium will require a shutdown of the compute nodes of all SciNet systems (Niagara, Mist, Rouge, Teach, as well as hosted equipment) from 8 AM to 5 PM EST. The login nodes, file systems and the HPSS system will remain available. The scheduler will hold jobs that are submitted until the maintenance has finished.&lt;br /&gt;
&lt;br /&gt;
'''January 9, 2025 11:00 am EST:''' Systems are back online&lt;br /&gt;
&lt;br /&gt;
'''January 8, 2025 10:34 pm EST:''' We had some sort of thermal event at the datacenter, and the clusters are down. We're still investigating&lt;br /&gt;
&lt;br /&gt;
'''January 8, 2025 08:00 am EST:''' Balam, Rouge and Neptune are shutdown for electrical upgrades&lt;br /&gt;
&lt;br /&gt;
'''January 6, 2025:''' As part of the installation of the new computing cluster Trillium, there will be a (permanent) reduction in computing capacity of Niagara and Mist. Only 50% of Niagara and 35% of Mist will remain active after January 6th.  The reduction will require Mist to be shutdown for a few hours on January 6th. Balam, Rouge and Neptune will be shutdown on Wednesday January 8th for the same reason.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
'''December 20, 2024 09:00 am EST:''' OpenOnDemand service will not be available on Dec 20 from 9 a.m. to 5 p.m. due to scheduled maintenance.&lt;br /&gt;
&lt;br /&gt;
'''December 16, 2024, 08:21 am EST:''' The Niagara scheduler has been restarted.&lt;br /&gt;
  &lt;br /&gt;
'''December 16, 2024, 00:04 am EST:''' The Niagara scheduler has an issue; we are investigating.&lt;br /&gt;
  &lt;br /&gt;
'''Fri Nov 8, 2024, 09:45 AM EST.''' Balam and Rouge schedulers are back online.&lt;br /&gt;
&lt;br /&gt;
'''Thu Nov 7, 2024, 10:30 PM EST.''' Most systems are up, except for the schedulers on Balam and Rouge (but even their login nodes are up), and a few 'neptune' niagara nodes.&lt;br /&gt;
&lt;br /&gt;
'''Thu Nov 7, 2024, 5:30 PM EST:''' Systems are being brought up, but not yet available for users.&lt;br /&gt;
&lt;br /&gt;
'''Downtime Announcement: On Thu Nov 7, 2024, all systems and storage located at the SciNet Datacenter (Niagara, Mist, HPSS, Rouge, Teach, JupyterHub, Balam) will be unavailable from 7 a.m. to 5 p.m. ET.&lt;br /&gt;
This outage is required to install new electrical equipment (UPS) for the upcoming systems refresh. The work is expected to be completed in one day.&lt;br /&gt;
The scheduler will hold jobs that cannot finish before the start of the shutdown. Users are encouraged to submit small and short jobs that can take advantage of this, as the scheduler may be able to fit these jobs in before the maintenance on otherwise idle nodes.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 24 15:05 EDT 2024''': Cooling pump motor has been replaced. All systems are back to normal.&lt;br /&gt;
&lt;br /&gt;
'''Tue Oct 22 16:35 EDT 2024''': The motor is scheduled for replacement on Thursday, Oct 24.&lt;br /&gt;
&lt;br /&gt;
'''Mon Oct 21 17:15 EDT 2024''': Compute nodes will remain down until we can replace the main cooling pump.  This may take several days.  Please see this page for updates.&lt;br /&gt;
&lt;br /&gt;
'''Mon Oct 21 12:15 EDT 2024''': Compute nodes have been shutdown due to a cooling system failure.&lt;br /&gt;
&lt;br /&gt;
'''Fri Oct 18 21:40 EDT 2024''': Systems are back to normal&lt;br /&gt;
&lt;br /&gt;
'''Fri Oct 18 21:15 EDT 2024''': We are experiences technical difficulties, apparently caused by a glitch in the file systems&lt;br /&gt;
&lt;br /&gt;
'''Tue Oct 1 10:45 EDT 2024''': The Jupyter Hub service will be rebooted today at around 11:00 am EDT for system upgrades. &lt;br /&gt;
&lt;br /&gt;
'''Tue Sep 3 07:00 EDT 2024''': Intermittent file system issues which may cause issues logging in.  We are in the process of resolving the issue.&lt;br /&gt;
&lt;br /&gt;
'''Sun Sep 1 00:01 - 04:00 EDT 2024''': Network maintenance may cause connection issues to the datacentre.&lt;br /&gt;
&lt;br /&gt;
'''Thu Aug 22 13:30:00 EDT 2024''': Chiller issue caused about 25% of Niagara compute nodes to go down; users should resubmit any affected jobs.&lt;br /&gt;
&lt;br /&gt;
'''Wed Aug 21 16:35:00 EDT 2024''': Maintenance finished; compute nodes are now available for user jobs.&lt;br /&gt;
&lt;br /&gt;
'''Wed Aug 21 7:00:00 EDT 2024''': Maintenance started.&lt;br /&gt;
&lt;br /&gt;
'''Sun Aug 18 19:15:00 EDT 2024''': Issues have been resolved.&lt;br /&gt;
&lt;br /&gt;
'''Sun Aug 18 14:30:00 EDT 2024''': Power issues seem to have brought compute nodes down, and compounded to the file system issues we had earlier.&lt;br /&gt;
&lt;br /&gt;
'''Sun Aug 18 10:31:53 EDT 2024''': GPFS is back online, and seems to be holding&lt;br /&gt;
&lt;br /&gt;
'''Sun Aug 18 08:44:40 EDT 2024''': Sorry, problems with GPFS file systems are reoccurring. &lt;br /&gt;
&lt;br /&gt;
'''Sun Aug 18 07:59:02 EDT 2024''': GPFS file systems are back to normal. Many jobs have died and will need to be resubmitted.&lt;br /&gt;
&lt;br /&gt;
'''Sun Aug 18 06:39:12 EDT 2024''': Support staff detected the problem and started to work on the fix&lt;br /&gt;
&lt;br /&gt;
'''Sun Aug 18 00:53:52 EDT 2024''': GPFS file systems (home, scratch, project) started to show initial stages of problems&lt;br /&gt;
&lt;br /&gt;
'''August 21, 2024''': The annual cooling tower maintenance for the SciNet data centre will take place on August 21, 2024 from 7 a.m. EDT until the end of day. This maintenance requires a shutdown of the compute nodes of all SciNet systems (Niagara, Mist, Rouge, Teach, as well as hosted equipment). The login nodes, file systems and the HPSS system will remain available.&lt;br /&gt;
&lt;br /&gt;
The scheduler will hold jobs that cannot finish before the start of the shutdown. Users are encouraged to submit small and short jobs that can take advantage of this, as the scheduler may be able to fit these jobs in before the maintenance on otherwise idle nodes.&lt;br /&gt;
&lt;br /&gt;
'''Thursday, August 1, 10:00 PM EDT''' Filesystem problems resolved.&lt;br /&gt;
&lt;br /&gt;
'''Thursday, August 1, 9:30 PM EDT''' Filesystem problems preventing logins to the systems.  Working on it.&lt;br /&gt;
&lt;br /&gt;
'''Monday, July 22, 11:50 AM EDT''' Systems are back to normal&lt;br /&gt;
&lt;br /&gt;
'''Monday, July 22, 10:50 AM EDT''' Cooling problem has been fixed. Systems are coming up&lt;br /&gt;
&lt;br /&gt;
'''Monday, July 22, 10:20 AM EDT''' Compute nodes have been shutdown due to a cooling tower failure.&lt;br /&gt;
&lt;br /&gt;
'''Friday, July 19, 9:30 AM EDT''' CCEnv modules available on all login nodes again.&lt;br /&gt;
&lt;br /&gt;
'''Friday, July 19, 5:00 AM EDT''' Some login nodes do not have the CCEnv modules available.  We are working on a fix.&lt;br /&gt;
&lt;br /&gt;
'''Monday, Jun 3, 12:55 PM EDT''' All systems are recovered now.&lt;br /&gt;
&lt;br /&gt;
'''Monday, Jun 3, 10:50 AM EDT''' The file system issues affect all nodes, so all systems are inaccessible to users at the moment. No time estimate yet for when the systems may be back.&lt;br /&gt;
&lt;br /&gt;
'''Monday, Jun 3, 7:58 AM EDT''' Login issues for Niagara and Mist. There are file system issues as well. Investigating.&lt;br /&gt;
&lt;br /&gt;
'''Sunday, Jun 2, 12:00 PM EDT''' CCEnv modules missing, investigating.&lt;br /&gt;
&lt;br /&gt;
'''Wednesday May 29, 5:50 PM EDT''' Niagara compute nodes are up.  &lt;br /&gt;
&lt;br /&gt;
'''Wednesday May 29, 4:40 PM EDT''' Niagara compute nodes are coming up.  &lt;br /&gt;
&lt;br /&gt;
'''Wednesday May 29, 4 PM EDT''' Niagara login nodes and jupyterhub are up; file system is now accessible.  &lt;br /&gt;
&lt;br /&gt;
'''Wednesday May 29, 2 PM EDT''' Electricians are checking and testing all junction boxes and connectors under the raised floor for safety.  Some systems are expected to be back up later today (storage, login nodes), and compute systems will be powered up as soon as it is deemed safe.&lt;br /&gt;
&lt;br /&gt;
'''Tuesday May 28, 3 PM EDT''' Cleaning crews are at the datacentre, to pump the water and install dryers.  Once the floors are dry, we need to inspect all electrical boxes to ensure safety.  We do not expect to have a fully functional datacentre before Thursday, although we hope to be able to turn on the storage and login nodes sometime tomorrow, if circumstances permit.  Apologies, and thank you for your patience.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
'''Tuesday May 28, 7 AM EDT''' A water mains break outside our datacentre has caused extensive flooding, and all systems have been shut down preventatively. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
'''Friday May 17, 10 PM EDT - Saturday May 18, 2 AM EDT:''' The external network will be unavailable for maintenance. Running and queued jobs on the systems will not be affected.&lt;br /&gt;
&lt;br /&gt;
'''Tuesday May 14, 6:45 PM EDT:''' All systems are recovered now.&lt;br /&gt;
&lt;br /&gt;
'''Tuesday May 14, 5 PM EDT:''' Power loss at the datacentre resulted in loss of cooling.  Systems are being restored.&lt;br /&gt;
&lt;br /&gt;
'''Friday May 3, 10 PM EDT - Saturday May 4, 2 AM EDT:''' The external network will be unavailable for maintenance. Running and queued jobs on the systems will not be affected.&lt;br /&gt;
&lt;br /&gt;
'''Tuesday April 17, 2024: 11:00 ''' The restart of the Niagara login nodes has been completed successfully.&lt;br /&gt;
&lt;br /&gt;
'''Tuesday April 17, 2024: 09:40 ''' Niagara login nodes will be rebooted &lt;br /&gt;
&lt;br /&gt;
'''Tuesday April 16, 2024: 12:45 ''' mist-login01  recovered now&lt;br /&gt;
&lt;br /&gt;
'''Tuesday April 16, 2024: 11:45 ''' mist-login01  will be unavailable due to maintenance from 12:15 to 12:45. Following the completion of maintenance, login access should be restored &lt;br /&gt;
&lt;br /&gt;
'''Monday April 15, 2024: 13:02 ''' Balam-login01 will be unavailable due to maintenance from 13:00 to 13:30. Following the completion of maintenance, login access should be restored and available once more. &lt;br /&gt;
&lt;br /&gt;
'''Monday March 18, 2024: 14:45 ''' File system issue resolved.  Users are advised to check if their running jobs were affected, and if so, to resubmit.&lt;br /&gt;
&lt;br /&gt;
'''Monday March 18, 2024: 13:02 ''' File system issues.  This affects the ability to log in. We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''Monday March 11, 2024: 14:05 ''' All systems are recovered now&lt;br /&gt;
&lt;br /&gt;
'''Monday March 11, 2024:''' There will be an shutdown of the file system at SciNet for an emergency repair. As a consequence, the login nodes and compute nodes of all SciNet clusters using the file system (Niagara, Mist, Balam, Rouge, and Teach) will be down from 11 am EST until later in the afternoon. &lt;br /&gt;
&lt;br /&gt;
'''February 28, 2024, 16:30 PM EDT:''' All systems are recovered now.&lt;br /&gt;
&lt;br /&gt;
'''February 28, 2024, 1:00 PM EDT:''' A loop pump fault caused many compute nodes overheat. If you jobs failed around this time, please resubmit. Once the root cause has been addressed, the cluster will be brought up completely. Please report issues to support@scinet.utoronto.ca.&lt;br /&gt;
&lt;br /&gt;
'''February 22, 2024, 5:45 PM EDT:''' Maintenance finished and system restored. Please report issues to support@scinet.utoronto.ca.&lt;br /&gt;
&lt;br /&gt;
'''February 21, 2024, 7:00 AM EDT:''' Maintenance starting.  Niagara login nodes and the file system are kept up as much as possible, but will be rebooted at some point.&lt;br /&gt;
&lt;br /&gt;
'''February 20, 2024, 3:45 PM EDT:''' Cooling tower has been restored, all systems are in production. &lt;br /&gt;
&lt;br /&gt;
'''February 20, 2024, 1:30 AM EDT:''' Cooling tower malfunction, all compute nodes are shutdown, the root cause will be addressed earliest in the morning.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;span style=&amp;quot;color:red&amp;quot;&amp;gt;&amp;lt;b&amp;gt; February 21 and 22, 2024: SciNet Data Centre Maintenance:&amp;lt;/b&amp;gt;&amp;lt;/span&amp;gt;&amp;lt;br/&amp;gt;&lt;br /&gt;
This annual winter maintenance involves a full data centre shutdown&lt;br /&gt;
starting at 7:00 am EST on Wednesday, February 21st.  None of the&lt;br /&gt;
SciNet systems (Niagara, Mist, Rouge, Teach, the file systems, as&lt;br /&gt;
well as hosted equipment) will be accessible.  All systems should be&lt;br /&gt;
fully available again in the last afternoon of the 22nd.&lt;br /&gt;
&lt;br /&gt;
The scheduler will hold jobs that cannot finish before the start of&lt;br /&gt;
the shutdown. Users are encouraged to submit small and short jobs&lt;br /&gt;
that can take advantage of this, as the scheduler may be able to fit&lt;br /&gt;
these jobs in before the maintenance on otherwise idle nodes.&lt;br /&gt;
&lt;br /&gt;
'''Mon January 29, 08:20    (EST):''' Access to Niagara login nodes restored (it was an internal routing issue).&lt;br /&gt;
&lt;br /&gt;
'''Mon January 29, 07:35    (EST):''' No access to Niagara login nodes.  We are investigating.  Use the Mist login to get access to SciNet systems.&lt;br /&gt;
&lt;br /&gt;
'''Wed January 24, 15:20    (EST):''' maintenance on rouge-login01 &lt;br /&gt;
&lt;br /&gt;
'''Wed January 24, 14:55    (EST):''' Rebooting rouge-login01 &lt;br /&gt;
&lt;br /&gt;
'''Tue January 23, 10:25 am (EST):''' Mist-login01 maintenance done &lt;br /&gt;
&lt;br /&gt;
'''Tue January 23, 10:10 am (EST):''' Rebooting Mist-login01 to deploy new image&lt;br /&gt;
&lt;br /&gt;
'''Tue January 22, 21:00 am (EST):''' HPSS performance for hsi &amp;amp; htar clients is back to normal.&lt;br /&gt;
&lt;br /&gt;
'''Tue January 20, 11:50 am (EST):''' HPSS hsi/htar/VFS jobs will remain on PD state on the queue over the weekend, so we may work on archive02/vfs02 on Monday, and try to improve transfer performance. In the meantime you may use Globus (computecanada#hpss) if your workflow is suitable. &lt;br /&gt;
&lt;br /&gt;
'''Tue January 14, 13:20 am (EST):''' The ongoing HPSS jobs from Friday finished earlier, so we restarted HPSS sooner and released the PD jobs on the queue. &lt;br /&gt;
&lt;br /&gt;
'''Tue January 12, 10:40 am (EST):''' We have applied some tweaks to the HPSS configuration to improve performance, but they won't take effect until we restart the services, which scheduled for Monday morning. If over the weekend we notice that there are no HPSS jobs running on the queue we may restart HPSS sooner. &lt;br /&gt;
&lt;br /&gt;
'''Tue January 09, 9:10 am (EST):''' Remaining cvmfs issues cleared.&lt;br /&gt;
&lt;br /&gt;
'''Tue January 09, 8:00 am (EST):''' We're investigating remaining issues with cvmfs access on login nodes.&lt;br /&gt;
&lt;br /&gt;
'''Mon January 08, 21:50 pm (EST):''' File systems are back to normal. Please resubmit your jobs.  &lt;br /&gt;
&lt;br /&gt;
'''Mon January 08, 9:10 pm (EST):''' We had a severe deadlock, and some disk volumes went down. The file systems are being recovered now. It could take another hour.&lt;br /&gt;
&lt;br /&gt;
'''Mon January 08, 7:20 pm (EST):''' We seem to have a problem with the file system, and are investigating.&lt;br /&gt;
&lt;br /&gt;
'''Tue December 19, 2:45 pm (EST):''' Compute nodes are available again.  &lt;br /&gt;
&lt;br /&gt;
'''Tue December 19, 12:09 pm (EST):''' Maintenance was postponed by one hour. &lt;br /&gt;
&lt;br /&gt;
'''Tue December 19, 12 noon - 1 pm (EST):''' There will be a shutdown of the compute nodes of the Niagara, Mist and Rouge cluster to allow for an emergency repair to the cooling tower.  Login nodes will remain available but no jobs will run during that time.  Updates will be posted on here.&lt;br /&gt;
&lt;br /&gt;
'''Mon Dec  11 11:17:00 EST 2023:''' File systems recovered; Niagara and Mist are operational again.&lt;br /&gt;
&lt;br /&gt;
'''Mon Dec  11 7:51:00 EST 2023:''' Niagara's login nodes are being overwhelmed.  We are investigating. Likely file-system related.&lt;br /&gt;
&lt;br /&gt;
'''Thu Dec  6 10:01:24 EST 2023:''' Niagara's scheduler rebooting for security patches.&lt;br /&gt;
&lt;br /&gt;
'''Wed Dec  6 13:06:46 EST 2023:''' Endpoint computecanada#niagara transition from Globus GCSv4 to GCSv5 is completed. computecanada#niagara-GCSv4 has been deactivated&lt;br /&gt;
&lt;br /&gt;
'''Mon Dec  4 16:35:07 EST 2023:''' Endpoint computecanada#niagara has now been upgraded to Globus GCSv5. The old endpoint is still available as computecanada#niagara-GCSv4 on nia-datamover2, only until Wednesday, at which time we'll disable it as well.&lt;br /&gt;
&lt;br /&gt;
'''Mon Dec  4 11:54:49 EST 2023:''' The nia-datamover1 node will the offline this Monday afternoon for the Globus GCSv5 upgrade. Endpoint computecanada#niagara-GCSv4 will still be available via nia-datamover2&lt;br /&gt;
&lt;br /&gt;
'''Tue Nov 28 16:29:14 EST 2023:''' The computecanada#hpss Globus endpoint is now running GCSv5. We'll find a window of opportunity next week to upgrade computecanada#niagara to GCSv5 as well.&lt;br /&gt;
&lt;br /&gt;
'''Tue Nov 28 14:20:30 EST 2023:''' The computecanada#hpss Globus endpoint will be offline for the next few hours for the GCSv5 upgrade.&lt;br /&gt;
&lt;br /&gt;
'''Fri Nov 10, 2023, 18:00 PM EDT:''' The HPSS upgrade is finished. We didn't have time to update Globus to GCSv5, so we'll find a window of opportunity to do this next week. &lt;br /&gt;
&lt;br /&gt;
Please be advised that starting this &amp;lt;B&amp;gt;Friday morning, Nov/10, we'll be upgrading the HPSS system from version 8.3 to 9.3 and the HPSS Globus server from GCSv4 to GCSv5.&amp;lt;/B&amp;gt; Everything going well we expect to be back online by the end of the day.  &lt;br /&gt;
&lt;br /&gt;
'''Fri Nov 3, 2023, 12:20 PM EDT:''' The &amp;quot;Niagara at Scale&amp;quot; event has finished. Niagara is available again for all users.&lt;br /&gt;
&lt;br /&gt;
'''Tue Oct 31, 2023, 12 PM EDT:''' The &amp;quot;Niagara at Scale&amp;quot; event has started.&lt;br /&gt;
&lt;br /&gt;
'''Tue Oct 31, 2023, 12:PM EDT - Fri Nov 3, 2023, 12:00 PM EDT:''' Three-day reservation for the &amp;quot;Niagara at Scale&amp;quot; event. Only &amp;quot;Niagara at Scale&amp;quot; projects will run on the compute nodes. Users are encouraged to submit small and short jobs that could run before this event.  Throughout the event, users can still login, access their data, and submit jobs, but these jobs will not run until after the event. Note that the debugjob queue will remain available to everyone as well.&lt;br /&gt;
&lt;br /&gt;
''' Thu Oct 27 11:16 AM EDT:''' SSH keys are gradually being restored, estimated to complete by 1:15 PM.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 27, 2023, 8:00 EDT:''' SSH key login authentication with CCDB keys is currently not working, on many Alliance systems.  It appears this started last night. Issue is being investigated.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 26, 2023, 12:35 EDT:''' Mist login node is accessible again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 26, 2023, 12:05 EDT:''' Mist login node is under maintenance and temporarily inaccessible to users.&lt;br /&gt;
&lt;br /&gt;
'''Wed Oct 25 7:54 PM EDT:''' slurm-*.out now outputs job info for last array job.&lt;br /&gt;
&lt;br /&gt;
'''Tue Oct 24 12:00 AM EDT:''' network appears to be up&lt;br /&gt;
&lt;br /&gt;
'''Tue Oct 24 11:32 AM EDT:''' campus network issues&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 05, 2023, 12:05 PM EDT:''' Niagara scheduler is back online.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 05, 2023, 11:50 AM EDT:''' Niagara scheduler is temporarily under maintenance for security updates. &lt;br /&gt;
&lt;br /&gt;
''' Thu Sep 28, 2023 11:00 am''': Niagara scheduler is back online.&lt;br /&gt;
&lt;br /&gt;
''' Thu Sep 28, 2023 10:50 am''': Niagara scheduler is temporarily under maintenance for security updates.&lt;br /&gt;
&lt;br /&gt;
''' Wed Sep 27, 2023 11:35 am''': Mist login node is accessible again.&lt;br /&gt;
&lt;br /&gt;
''' Wed Sep 27, 2023 11:00 am''': Mist login node is under maintenance and temporarily inaccessible to users.&lt;br /&gt;
&lt;br /&gt;
''' Wed Sep 6, 2023 11:30 am''': Mist login node is accessible again.&lt;br /&gt;
&lt;br /&gt;
''' Wed Sep 6, 2023 11:00 am''': Mist login node is under maintenance and temporarily inaccessible to users.&lt;br /&gt;
&lt;br /&gt;
''' Fri Aug 25, 2023 0:19 am''': A power glitch brought some compute nodes down; users should resubmit any affected jobs. The Jupyterhub had to be restarted for the same reason.&lt;br /&gt;
&lt;br /&gt;
''' Mon Aug 14, 2023 12:10 pm''': Network problems with Teach cluster are now resolved and it is again available for users.&lt;br /&gt;
&lt;br /&gt;
''' Mon Aug 14, 2023 11:40 am''': Network problems with Teach cluster. We are investigating.&lt;br /&gt;
&lt;br /&gt;
''' Thu Aug 3, 2023 11:10 am''': Mist login node is accessible again.&lt;br /&gt;
&lt;br /&gt;
''' Thu Aug 3, 2023 10:40 am''': Mist login node is under maintenance and temporarily inaccessible to users.&lt;br /&gt;
&lt;br /&gt;
''' Tue Aug 1, 2023 2:43 pm''': To recover from the power glitch, all servers on the SciNet jupyterhub have been stopped. Please restart you server if you need to.&lt;br /&gt;
&lt;br /&gt;
''' Tue Aug 1, 2023 11:46 am''': There was a power glitch at 11:46 Aug 1, 2023, causing a significant number of job losses. Please resubmit your jobs.&lt;br /&gt;
&lt;br /&gt;
'''Summer Maintenance Shutdown Finished''' -- Slurm upgraded to version 23.02.3.&lt;br /&gt;
Change to be aware: SLURM_NTASKS is only set if --ntasks option is set.&lt;br /&gt;
Details at: https://bugs.schedmd.com/show_bug.cgi?id=17108&lt;br /&gt;
&lt;br /&gt;
'''July 17 and 18, 2023''':  Announcement: Summer Maintenance Shutdown&amp;lt;/span&amp;gt;&amp;lt;br&amp;gt; &lt;br /&gt;
&lt;br /&gt;
'''July 17th, 2023''' This maintenance involves a full data centre shutdown will start at 7:00 a.m. ET on Monday July 17th, 2023. None of the SciNet systems (Niagara, Mist, Rouge, Teach, the file systems, as well as hosted equipment) will be accessible.&lt;br /&gt;
&lt;br /&gt;
'''July 18th, 2023''' The shutdown will last until Tuesday July 18th, 2023. Systems are expected to be fully available in the evening of that day.&lt;br /&gt;
&lt;br /&gt;
The scheduler will hold jobs that cannot finish before the start of the shutdown. Users are encouraged to submit small and short jobs that can take advantage of this, as the scheduler may be able to fit these jobs in before the maintenance on otherwise idle nodes.&lt;br /&gt;
&lt;br /&gt;
'''Wed Jun 21 16:03:45 EDT 2023:''' Niagara's scheduler maintenance is finished.&lt;br /&gt;
&lt;br /&gt;
'''Wed Jun 21 15:42:00 EDT 2023:''' Niagara's scheduler is rebooting in 10 minutes for a short maintenance down time.&lt;br /&gt;
&lt;br /&gt;
'''Wed Jun 21, 2023, 11:25 AM EDT:''' Maintenance is finished and Teach cluster is accessible again.&lt;br /&gt;
&lt;br /&gt;
'''Tue Jun 20, 2023, 9:55 AM EDT:''' Teach cluster is powered off for maintenance.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;span style='color:red'&amp;gt;'''Tue June 20, 2023:'''  Announcement:&amp;lt;/span&amp;gt;&amp;lt;br&amp;gt; The Teach cluster at SciNet will undergo a maintenance shutdown starting on Tuesday June 20, 2023.  It will likely take a few days before it will be available again.  Check here for updates.&lt;br /&gt;
&lt;br /&gt;
'''Mon Jun 5, 2023, 2:35 PM EDT:''' All systems are operational again.&lt;br /&gt;
&lt;br /&gt;
'''Mon Jun 5, 2023, 11:55 AM EDT:''' There were issues with the cooling system.  The login nodes and file systems are now accessible again, but compute nodes are still off.&lt;br /&gt;
&lt;br /&gt;
'''Mon Jun 5, 2023, 6:55 AM EDT:''' Issues at the data center, we are investigating.&lt;br /&gt;
&lt;br /&gt;
'''Sat May 27, 2023, 21:00AM EDT:''' We have been able to mitigate the UPS issue for now, until new parts arrive sometime during the week. System will be accessible soon&lt;br /&gt;
&lt;br /&gt;
'''Sat May 27, 2023, 16:00AM EDT:''' We identified an UPS/Power related issue on the datacenter, that is adversely affecting several components, in particular all file systems. Out of an abundance of caution we are shutting down the cluster, until the UPS situation is resolved. Ongoing jobs will be canceled.&lt;br /&gt;
&lt;br /&gt;
'''Sat May 27, 2023, 11:18AM EDT:''' Filesystem issues, investigating.&lt;br /&gt;
&lt;br /&gt;
'''Wed May 24, 2023, 11:40AM EDT:''' Mist login node is accessible again.&lt;br /&gt;
&lt;br /&gt;
'''Wed May 24, 2023, 11:10 AM EDT:''' Mist login node is under maintenance and temporarily inaccessible to users.&lt;br /&gt;
&lt;br /&gt;
'''Mon May 15, 2023, 10:08 AM EDT''' rebooting Mist-login node again &lt;br /&gt;
&lt;br /&gt;
'''Mon May 15, 2023, 09:15 AM EDT''' rebooting Mist-login node&lt;br /&gt;
&lt;br /&gt;
'''Mon May 01, 2023, 04:00 PM EDT''' done rebooting nia-login nodes&lt;br /&gt;
&lt;br /&gt;
'''Mon May 01, 2023, 12:00 PM EDT''' rebooting all nia-login nodes one at a time &lt;br /&gt;
&lt;br /&gt;
'''Mon May 01, 2023, 11:00 AM EDT''' nia-login07 is going to be rebooted.&lt;br /&gt;
&lt;br /&gt;
'''Thu Apr 20, 2023, 12:05 PM EDT:''' Mist login node is accessible again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Apr 20, 2023, 11:30 AM EDT:''' Mist login node is under maintenance and temporarily inaccessible to users.&lt;br /&gt;
&lt;br /&gt;
'''Thu Apr 20, 2023, 8:27 AM EDT:''' Intermittent file system issues. We are investigating.  For now (10:45 AM), the file systems appear operational.&lt;br /&gt;
&lt;br /&gt;
'''Fri 14 Apr 2023 10:25 AM EDT:''' Switch problem resolved.&lt;br /&gt;
&lt;br /&gt;
'''Fri 14 Apr 2023 10:10 AM EDT:''' A switch problem is affecting access to certain equipment at the SciNet data center, including the Teach cluster.  Niagara and Mist are accessible.&lt;br /&gt;
&lt;br /&gt;
'''Fri 14 Apr 2023 09:55 AM EDT:''' SciNet Jupyter Hub maintenance is finished and it is again available for users.&lt;br /&gt;
&lt;br /&gt;
'''Fri 14 Apr 2023:''' SciNet Jupyter Hub will be restarted for system updates this morning.  Keep in mind to save your notebooks!&lt;br /&gt;
&lt;br /&gt;
'''Thu 06 Apr 2023 03:40 PM EDT:''' Rouge cluster is accessible again.&lt;br /&gt;
&lt;br /&gt;
'''Thu 06 Apr 2023 01:00 PM EDT:''' Rouge cluster is temporarily inaccessible to users due to the electrical work.&lt;br /&gt;
&lt;br /&gt;
'''Sun 02 Apr 2023 03:37 AM EDT:''' IO/read errors on the file system seem to have been fixed. Please resubmit your jobs, and report any further problems to support. Burst Buffer will remain offline for now.&lt;br /&gt;
&lt;br /&gt;
'''Sun 02 Apr 2023 00:18 AM EDT:''' File System is back up, but there seems to be some IO/read errors. All running jobs have been killed. Please hold off on submitting jobs until further notice.&lt;br /&gt;
&lt;br /&gt;
'''Sat 01 Apr 2023 10:17 PM EDT:''' We are having issues with the File System. Currently investigating the cause.&lt;br /&gt;
&lt;br /&gt;
'''Fri 31 Mar 2023 11:00 PM EDT:''' Burst Buffer may be the culprit. We are investigating but may have to take Burst Buffer offline. &lt;br /&gt;
&lt;br /&gt;
'''Fri 31 Mar 2023 01:30 PM EDT:''' File system issues causing trouble for some jobs on Niagara and Mist&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Tue 28 Mar 2023 11:05 AM EDT: Mist login node is accessible again.&lt;br /&gt;
&lt;br /&gt;
Tue 28 Mar 2023 10:35 AM EDT: Mist login node is under maintenance and temporarily inaccessible to users.&lt;br /&gt;
&lt;br /&gt;
Fri 17 Mar 2023 14:50 PM EDT: All systems online.&lt;br /&gt;
&lt;br /&gt;
Fri 17 Mar 2023 11:00 AM EDT: Problem identified and repaired. Starting to bring up systems, but not available to users yet.&lt;br /&gt;
&lt;br /&gt;
Fri 17 Mar 2023 09:15:39 EDT: Staff on site and ticket opened with cooling contractor, cause of failure unclear &lt;br /&gt;
&lt;br /&gt;
Fri 17 Mar 2023 01:47:43 EDT: Cooling system malfunction, datacentre is shut down. &lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Tue Feb 28, 16:40 EST:&amp;lt;/b&amp;gt; All systems are back online.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Tue Feb 28, 15:30 EST:&amp;lt;/b&amp;gt; Maintenance is complete. Bringing up systems.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Tue Feb 28, 7:10 AM EST:&amp;lt;/b&amp;gt; Maintenance shutdown resuming.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Mon Feb 27, 3:55 PM EST:&amp;lt;/b&amp;gt; Maintenance paused as parts were delayed. The maintenance will resume tomorrow (Tue Feb 28) at 7AM EST for about 5 hours.  In the meantime, the login nodes of the systems will be brought online.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Mon Feb 27, 7:20 AM EST:&amp;lt;/b&amp;gt; Maintenance shutdown started.&lt;br /&gt;
 &lt;br /&gt;
&amp;lt;span style=&amp;quot;color:red&amp;quot;&amp;gt;&amp;lt;b&amp;gt; February 27 and 28, 2023: SciNet Data Centre Maintenance:&amp;lt;/b&amp;gt;&amp;lt;/span&amp;gt;&amp;lt;br/&amp;gt;&lt;br /&gt;
This annual winter maintenance involves a full data centre shutdown&lt;br /&gt;
starting at 7:00 a.m. EST on Monday, February 27. None of the SciNet&lt;br /&gt;
systems (Niagara, Mist, Rouge, Teach, the file systems, as well as&lt;br /&gt;
hosted equipment) will be accessible.&lt;br /&gt;
&lt;br /&gt;
On the second day of the maintenance, Niagara, Mist, and their file&lt;br /&gt;
systems are expected to become partially available for users.  All&lt;br /&gt;
systems should be fully available in the evening of the 28th.&lt;br /&gt;
&lt;br /&gt;
The scheduler will hold jobs that cannot finish before the start of&lt;br /&gt;
the shutdown. Users are encouraged to submit small and short jobs&lt;br /&gt;
that can take advantage of this, as the scheduler may be able to fit&lt;br /&gt;
these jobs in before the maintenance on otherwise idle nodes.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Feb 17, 2023, 11:15 PM EST:&amp;lt;/b&amp;gt; File system issues on Teach fixed and Teach is accessible again. Note that the file system of Teach is not very good at handling many remote vscode connections.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Feb 17, 2023, 11:02 PM EST:&amp;lt;/b&amp;gt; File system issues on Teach.  We are working on a fix.  &lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Sun Feb 12, 2023, 3:05 PM EST&amp;lt;/b&amp;gt; All systems are back online.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Sun Feb 12, 2023, 2:10 PM EST&amp;lt;/b&amp;gt; Powers restored, clusters are being started.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Sat Feb 11, 2023, 2:35 PM EST&amp;lt;/b&amp;gt; Powers interruption started. All compute nodes will be down, likely until Sunday &lt;br /&gt;
afternoon.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Sat Feb 11, 2023, 1:20 PM EST&amp;lt;/b&amp;gt; There is to be an emergency power repair on the adjacent street. The datacentre will be &lt;br /&gt;
switching over to generator. All compute nodes will be down.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Fri Feb 10, 2023, 10:55 AM EST&amp;lt;/b&amp;gt; All systems are back online.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Fri Feb 10, 2023, 10:00 AM EST&amp;lt;/b&amp;gt; Cooling issue resolved, cluster is being started.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Wed Jan 25, 2023, 02:15 PM EST&amp;lt;/b&amp;gt; Mist login node is accessible again.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Wed Jan 25, 2023, 10:30 AM EST&amp;lt;/b&amp;gt; Mist login node is under maintenance and temporarily inaccessible to users.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Mon Jan 23, 2023, around 7-8 AM EST&amp;lt;/b&amp;gt; Intermediate file system issuse may have killed your job. Users are advised to resubmit.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Sat Jan 21, 2023, 00:50 EST&amp;lt;/b&amp;gt; Niagara, Mist, Rouge and the filesystems are up&lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Fri Jan 20, 2023, 11:19 PM: EST&amp;lt;/b&amp;gt; Systems are coming up. We have determined that there was a general power glitch in the area of our Datacentre. The power has been fully restored&lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Fri Jan 20, 2023, 10:34 PM: EST&amp;lt;/b&amp;gt; Cooling is back. Systems are slowly coming up  &lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Fri Jan 20, 2023, 8:20 PM: EST&amp;lt;/b&amp;gt; A cooling failure at the data center, possibly due to a power glitch. We are investigating.  &lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Thu Jan 12, 2023, 9:30 AM EST&amp;lt;/b&amp;gt; File system is experiencing issues. Issues have stabilized, but jobs running around this time may have been affected.&lt;br /&gt;
&lt;br /&gt;
'''Wed Dec 21, 2022, 12:00 PM: ''' Please note that SciNet is on vacation, together with the University of Toronto. Full service will resume on Jan 2, 2023. We will endeavour to keep systems running, and answer tickets, on a best-effort basis.  Happy Holidays!!!&lt;br /&gt;
&lt;br /&gt;
'''Fri Dec 16, 2022, 2:19 PM: ''' City power glitch caused all compute nodes to reboot. Please resubmit your jobs.&lt;br /&gt;
&lt;br /&gt;
'''Mon Dec 12, 2022, 9:30 AM - 11:30:''' File system issues caused login issues and may have affected running jobs.  System back to normal now, but users may want to check any jobs they had running. &lt;br /&gt;
&lt;br /&gt;
'''Wed Dec 7, 2022, 11:40 AM EST:''' Systems are being brought back online.&lt;br /&gt;
&lt;br /&gt;
'''Wed Dec 7, 2022, 09:00 AM EST:''' Maintenance is underway.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;span style='color:red'&amp;gt;&amp;lt;b&amp;gt;Announcement:&amp;lt;/b&amp;gt;&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
On '''Wednesday December 7th, 2022''', the file systems of the SciNet's systems, Niagara, Mist, HPSS, Teach cluster, will undergo maintenance from 9:00 am EST.  During the maintenance, there will be no access to any of these systems, as it requires all file system operations to have stopped.  The maintenance should take about 1 hour, and all systems are expected to become available again later that morning.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
'''Wed Nov 30, 2022, 14:45 PM EST:''' Mist login node is accessible again.&lt;br /&gt;
&lt;br /&gt;
'''Wed Nov 30, 2022, 14:15 PM EST:''' Mist login node is under maintenance and temporarily inaccessible to users. &lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 20, 2022, 18:00 PM EDT:''' Systems are back online &lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 20, 2022, 09:40 AM EDT:''' About half of Niagara compute nodes are up. Note that only jobs that can finish by 5:00 PM will run.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 20, 2022, 07:50 AM EDT:''' Jupyter Hub is available again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 20, 2022, 07:35 AM EDT:''' Jupyter Hub is being updated and temporarily inaccessible to users.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 20, 2022, 07:30 AM EDT:''' Maintenance is underway.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;span style='color:red'&amp;gt;&amp;lt;b&amp;gt;Announcement:&amp;lt;/b&amp;gt;&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
On '''Thursday October 20th, 2022''', the SciNet datacentre (which hosts Niagara and Mist) will undergo transformer maintenance from 7:30 am EDT to 5:00 pm EDT.  At both the start and end of this maintenance window, all systems will need to be briefly shutdown and will not be accessible.  Apart from that, during this window, login nodes will be accessible and part of Niagara will be available to run jobs. The Mist and Rouge clusters will be off for the entirety of this maintenance. &lt;br /&gt;
&lt;br /&gt;
Users are encouraged to submit Niagara jobs of about 1 to 2 hours in the days before the maintenance, as these could be run within the&lt;br /&gt;
window of 8 AM and 5 PM EDT.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
'''Wed Oct 5, 2022, 12:10 PM EDT:''' A grid power glitch caused all compute nodes to reboot. Please resubmit your jobs.&lt;br /&gt;
&lt;br /&gt;
'''Mon Oct 3, 2022, 11:20 PM EDT:'''  Niagara login nodes are accessible from outside again.&lt;br /&gt;
&lt;br /&gt;
'''Mon Oct 3, 2022, 9:20 PM EDT:'''  Niagara login nodes are inaccessible from outside of the datacentre at the moment. As a work-around, ssh into mist.scinet.utoronto.ca and then ssh into e.g. nia-login01.&lt;br /&gt;
&lt;br /&gt;
'''Wed Sep 28, 2022, 1:15 PM EDT:''' The JupyterHub maintenance is finished and it is now accessible to users.&lt;br /&gt;
&lt;br /&gt;
'''Wed Sep 28, 2022, 1:00 PM EDT:''' The JupyterHub is to be rebooted for system upgrades. Running processes and notebooks will be closed. The service is expected to be back around 1:30 PM EDT.&lt;br /&gt;
 &lt;br /&gt;
'''Tue Sep 27, 2022, 11:50 AM EDT:''' Mist login node is accessible again.&lt;br /&gt;
&lt;br /&gt;
'''Tue Sep 27, 2022, 11:25 AM EDT:''' Mist login node is under maintenance and temporarily inaccessible to users.&lt;br /&gt;
&lt;br /&gt;
'''Mon Sep 26, 2022, 11:35 AM EDT:''' Rouge and Teach login nodes are accessible again.&lt;br /&gt;
&lt;br /&gt;
'''Mon Sep 26, 2022, 11:05 AM EDT:''' Rouge and Teach login nodes are under maintenance and temporarily inaccessible to users.&lt;br /&gt;
&lt;br /&gt;
'''Fri Sep 22, 2022, 0:46 AM EDT:''' The CCEnv software stack is back to normal.&lt;br /&gt;
&lt;br /&gt;
'''Thu Sep 22, 2022, 8:15 PM EDT:''' The CCEnv software stack is inaccessible due to an issue with CVMFS.&lt;br /&gt;
 &lt;br /&gt;
'''Tue Sep 20, 2022, 16:00 AM EDT:''' Rouge login node is accessible again.&lt;br /&gt;
&lt;br /&gt;
'''Tue Sep 20, 2022, 10:20 AM EDT:''' Rouge login node is under maintenance and temporarily inaccessible to users (hardware upgrade).&lt;br /&gt;
&lt;br /&gt;
'''Tue Sep 20, 2022, 9:41 AM EDT:''' Rouge login node is back up.&lt;br /&gt;
&lt;br /&gt;
'''Tue Sep 20, 2022, 8:25 AM EDT:''' Rouge login node down, we are investigating.&lt;br /&gt;
&lt;br /&gt;
'''Fri Sept 16, 2022, 9:30 AM EDT:''' Login nodes are accessible again.&lt;br /&gt;
&lt;br /&gt;
'''Fri Sept 16, 2022, 9:00 AM EDT:''' Login nodes are not accessible.  We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''Tue Sep 13, 2022, 11:00 AM EDT:''' Mist login node is available again.&lt;br /&gt;
&lt;br /&gt;
'''Tue Sep 13, 2022, 10:00 AM EDT:''' Mist login node is under maintenance and temporarily inaccessible to users.&lt;br /&gt;
&lt;br /&gt;
'''Fri Sep 2, 2022, 11:25 AM EDT:''' Rouge login node is back up.&lt;br /&gt;
&lt;br /&gt;
'''Fri Sep 2, 2022, 10:25 AM EDT:''' Issues with the Rouge login node; we are investigating.&lt;br /&gt;
&lt;br /&gt;
'''Tue Aug 23, 2022, 1:15 PM EDT:''' Jupyter Hub is available again.&lt;br /&gt;
&lt;br /&gt;
'''Tue Aug 23, 2022, 1:00 PM EDT:''' Jupyter Hub is being updated and temporarily inaccessible to users.&lt;br /&gt;
&lt;br /&gt;
'''Fri Aug 12, 2022, 6:30 PM EDT:''' File system issues are resolved.&lt;br /&gt;
&lt;br /&gt;
'''Fri Aug 12, 2022, 5:06 PM EDT:''' File system issues. We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''Thu Aug 11, 2022, 9:20 AM EDT:''' The login node issues have been resolved.&lt;br /&gt;
&lt;br /&gt;
'''Thu Aug 11, 2022, 7:50 AM EDT:''' We are having problems accessing the Niagara login nodes.  Until fixed, please login to Mist and then ssh to a Niagara login node to access Niagara (&amp;quot;ssh nia-login02&amp;quot;, for example).&lt;br /&gt;
&lt;br /&gt;
'''Fri July 15, 2022, 10:50 AM EDT:''' Jupyter Hub is available again.&lt;br /&gt;
&lt;br /&gt;
'''Fri July 15, 2022, 10:30 AM EDT:''' Jupyter Hub is being updated and temporarily inaccessible to users.&lt;br /&gt;
&lt;br /&gt;
'''Wed June 16, 2022, 3:45 PM EDT:''' File system is stable now. We're gradually opening the systems up.&lt;br /&gt;
&lt;br /&gt;
'''Wed June 16, 2022, 10:15 AM EDT:''' Emergency maintenance shutdown of filesystem. Running jobs will be affected.&lt;br /&gt;
&lt;br /&gt;
'''Wed June 15, 2022, 7:35 PM EDT:''' Maintenance shutdown finished. Most systems are available again.&lt;br /&gt;
&lt;br /&gt;
'''Wed June 15, 2022, 7:00 AM EDT:''' Maintenance shutdown of the SciNet datacentre. There will be no access to any of the SciNet systems during this time. We expect to be able to bring the systems back online in the evening of June 15th.&lt;br /&gt;
&lt;br /&gt;
'''Mon June 13, 2022, 7:00 AM EDT - Wed June 15, 2022, 7:00 AM EDT:''' Two-day reservation for the &amp;quot;Niagara at Scale&amp;quot; event. Only &amp;quot;Niagara at Scale&amp;quot; projects will run on the compute nodes (as well as SOSCIP projects, on a subset of nodes). Users are encouraged to submit small and short jobs that could run before this event.  Throughout the event, users can still login, access their data, and submit jobs, but these jobs will not run until after the subsequent maintenance (see below). Note that the debugjob queue will remain available to everyone as well.&lt;br /&gt;
&lt;br /&gt;
'''Mon May 30th, 2022, 12:42:00 EDT:''' Mist login node is available again.&lt;br /&gt;
&lt;br /&gt;
'''Mon May 30th, 2022, 10:22:00 EDT:''' Mist login node is being upgraded and temporarily inaccessible to users.&lt;br /&gt;
&lt;br /&gt;
'''Wed May 25th, 2022, 13:30:00 EDT:''' Niagara operating at 100% again.&lt;br /&gt;
&lt;br /&gt;
'''Tue May 24th, 2022, 21:30:00 EDT:''' Jupyter Hub up.  Part of Niagara can run compute jobs.&lt;br /&gt;
&lt;br /&gt;
'''Tue May 24th, 2022, 19:00:00 EDT:''' Systems are up. Users can login, BUT cannot submit jobs yet.&lt;br /&gt;
&lt;br /&gt;
'''Tue May 24th, 2022, 10:00:00 EDT:''' We are still performing system checks.&lt;br /&gt;
&lt;br /&gt;
'''Mon May 23rd, 2022, 16:44:30 EDT:''' Systems still down. Filesystems are working, but there are quite a number of drive failures - no data loss - so out of an abundance of caution we are keeping the systems down at least until tomorrow.  The long weekend has also been disruptive for service response, and we prefer to err on the safe side.&lt;br /&gt;
&lt;br /&gt;
'''Mon May 23rd, 2022, 08:12:14 EDT:''' Systems still down. Filesystems being checked to ensure no heat damage.&lt;br /&gt;
&lt;br /&gt;
'''Sun May 22nd, 2022, 10.16 am EDT:''' Electrician dispatched to replace blown fuses.&lt;br /&gt;
&lt;br /&gt;
'''Sun May 22nd, 2022, 2:54 am EDT:''' Automatic shutdown down due to power/cooling.&lt;br /&gt;
&lt;br /&gt;
'''Fri May 6th, 2022, 11:35 am EDT:''' HPSS scheduler upgrade also finished.&lt;br /&gt;
&lt;br /&gt;
'''Thu May 5th, 2022, 7:45 pm EDT:''' Upgrade of the scheduler has finished, with the exception of HPSS.&lt;br /&gt;
&lt;br /&gt;
'''Thu May 5th, 2022, 7:00 am - 3:00 pm EDT (approx):''' Starting from 7:00 am EDT, an upgrade of the scheduler of the Niagara, Mist, and Rouge clusters will be applied.  This requires the scheduler to be down for about 5-6 hours, and all compute and login nodes to be rebooted.&lt;br /&gt;
Jobs cannot be submitted during this maintenance, but jobs submitted beforehand will remain in the queue.  For most of the time, the login nodes of the clusters will be available so that users may access their files on the home, scratch, and project file systems.&lt;br /&gt;
&lt;br /&gt;
'''Monday May 2nd, 2022, 9:30 - 11:00 am EDT:''' the Niagara login nodes, the jupyter hub, and nia-datamover2 will get rebooted for updates.  In the process, any login sessions will get disconnected, and servers on the jupyterhub will stop. Jobs in the Niagara queue will not be affected.&lt;br /&gt;
&lt;br /&gt;
'''Tue Apr 26, 11:20 AM EDT:''' A Rolling update of the Mist cluster is taking a bit longer than expected, affecting logins to Mist. &lt;br /&gt;
 &lt;br /&gt;
'''Announcement:''' On Thursday April 14th, 2022, the connectivity to the SciNet datacentre will be disrupted at 11:00 AM EDT  for a few minutes, in order to deploy a new network core switch.  Any SSH connections or data transfers to SciNet systems (Niagara, Mist, etc.) may be terminated at that time.&lt;br /&gt;
&lt;br /&gt;
'''Thu March 24, 6:54 AM EST:''' HPSS is back online&lt;br /&gt;
&lt;br /&gt;
'''Thu March 24, 8:15 AM EST:''' HPSS has a hardware problem&lt;br /&gt;
&lt;br /&gt;
'''Wed March 2, 4:50 PM EST:''' The CCEnv software stack is available again on Niagara.&lt;br /&gt;
&lt;br /&gt;
'''Wed March 2, 7:50 AM EST:''' The CCEnv software stack on Niagara has issues; we are investigating.&lt;br /&gt;
 &lt;br /&gt;
'''Sat Feb 12 2022, 12:59 EST:''' Jupyterhub is back up, but may have hardware issue.&lt;br /&gt;
&lt;br /&gt;
'''Sat Feb 12 2022, 10:36 EST:''' Issue with the Jupyterhub, since last night.  We're investigating.&lt;br /&gt;
&lt;br /&gt;
'''Tue Feb 1 2022 19:20 EST:''' Maintenance finished successfully. Systems are up. &lt;br /&gt;
&lt;br /&gt;
'''Tue Feb 1 2022 13:00 EST:''' Maintenance downtime started.&lt;br /&gt;
&lt;br /&gt;
'''Mon Jan 31 2022 13:15:00 EST:''' The SciNet datacentre's cooling system needs an '''emergency repair''' as soon as possible.  During this repair, all systems hosted at SciNet (Niagara, Mist, Rouge, HPSS, and Teach) will need to be switched off and will be unavailable to users. Repairs will start '''Tuesday February 1st, at 1:00 pm EST''', and could take until the end of the next day.  Please check here for updates.&lt;br /&gt;
&lt;br /&gt;
'''Sat Jan 29 2020 16:45:38 EST:''' Fibre repaired.&lt;br /&gt;
&lt;br /&gt;
'''Sat 29 Jan 2022 11:22:27 EST:''' Fibre repair is underway.  Expect to have connectivity restored later today.&lt;br /&gt;
&lt;br /&gt;
'''Fri 28 Jan 2022 07:35:01 EST:''' The fibre optics cable that connects the SciNet datacentre was severed by uncoordinated digging at York University.  We expect repairs to happen as soon as possible.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 27 12:46 EST PM 2022:''' Network issues to and from the datacentre. We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''Sun Jan 23 11:05 EST AM 2022:''' Filesystem issues appear to have resolved.&lt;br /&gt;
&lt;br /&gt;
'''Sun Jan 23 10:30 EST AM 2022:''' Filesystem issues -- investigating.&lt;br /&gt;
&lt;br /&gt;
'''Sat Jan 8 11:42 EST AM 2022:''' The emergency maintenance is complete. Systems are up and available.&lt;br /&gt;
&lt;br /&gt;
'''Fri Jan 7 14:34 EST PM 2022:''' The SciNet shutdown is in progress. Systems are expected back on Saturday, Jan 8.&lt;br /&gt;
&lt;br /&gt;
'''&amp;lt;span style=&amp;quot;color:red&amp;quot;&amp;gt;Emergency shutdown Friday January 7, 2022&amp;lt;/span&amp;gt;''': An emergency shutdown of all SciNet to replace a crucial file system component is planned to take place on Friday January 7, 2022, starting at 8am EST, and will require at least 12 hours of downtime.  Updates will be posted during the day.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 6 08:20 EST AM 2022''' The SciNet filesystem is having issues.  We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''Fri Dec 24 13:31 EST PM 2021''' Please note the following scheduled network maintenance, which will result in loss of connectivity to the SciNet datacentre:  Start time&lt;br /&gt;
Dec 29, 00:30 EST  Estimated duration  4 hours and 30 minutes. &lt;br /&gt;
&lt;br /&gt;
'''Mon Dec 20 4:29 EST PM 2021''' Filesystem is back to normal. &lt;br /&gt;
&lt;br /&gt;
'''Mon Dec 20 2:53 EST PM 2021''' Filesystem problem - We are investigating. &lt;br /&gt;
&lt;br /&gt;
'''Wed Sep 23 12:30 EDT 2021 ''' Cooling restored.  Systems should be available later this afternoon.  &lt;br /&gt;
&lt;br /&gt;
'''Wed Sep 23 9:30 EDT 2021 ''' Technicians on site working on cooling system. &lt;br /&gt;
&lt;br /&gt;
'''Wed Sep 23 3:30 EDT 2021 ''' Cooling system issues still unresolved. &lt;br /&gt;
&lt;br /&gt;
'''Wed Sep 22 23:27:48 EDT 2021 ''' Shutdown of the datacenter due to a problem with the cooling system.&lt;br /&gt;
&lt;br /&gt;
'''Wed Sep 22 09:30 EDT 2021 ''': File system issues, resolved.&lt;br /&gt;
&lt;br /&gt;
'''Wed Sep 22 07:30 EDT 2021 ''': File system issues, investigating.&lt;br /&gt;
&lt;br /&gt;
'''Sun Sep 19 10:00 EDT 2021''': Power glitch interrupted all compute jobs; please resubmit any jobs you had running.&lt;br /&gt;
&lt;br /&gt;
'''Wed Sep 15 17:35 EDT 2021''': filesystem issues resolved&lt;br /&gt;
&lt;br /&gt;
'''Wed Sep 15 16:39 EDT 2021''': filesystem issues&lt;br /&gt;
&lt;br /&gt;
'''Mon Sep 13 13:15:07 EDT 2021''' HPSS is back online.&lt;br /&gt;
&lt;br /&gt;
'''Fri Sep 10 17:57:23 EDT 2021''' HPSS is offline due to unscheduled maintenance.&lt;br /&gt;
&lt;br /&gt;
'''Wed Aug 18 16:13:42 EDT 2021''' The HPSS upgrade is complete.&lt;br /&gt;
&lt;br /&gt;
'''HPSS Downtime August 17th and 18th, 2021 (Tuesday and Wednesday):''' We'll be upgrading the HPSS software to version 8.3, along with all the clients (htar/hsi, vfs and Globus/dsi)&lt;br /&gt;
&lt;br /&gt;
'''July 24, 2021, 6:00 PM EDT:''' There appear to be file system issues, which may affect users' ability to login.  We are investigating.&lt;br /&gt;
&lt;br /&gt;
''' July 23th, 2021, 9:00 AM EDT:''' ''' Security update: ''' Due to a severe vulnerability in the Linux kernel (CVE-2021-33909), our team is currently patching and rebooting all login nodes and compute nodes, as well as the JupyterHub.  There should be no affect on running jobs, however sessions on login and datamover nodes will be disrupted. &lt;br /&gt;
&lt;br /&gt;
''' July 20th, 2021, 7:00 PM EDT:''' ''' SLURM configuration''' - Changed the default behaviour to kill a job step if any task exits with a non-zero exit code. If your code is able to handle failures gracefully, please add srun's option --no-kill to recover the previous default behaviour.&lt;br /&gt;
&lt;br /&gt;
''' July 20th, 2021, 7:00 PM EDT:''' Maintenance finished, systems are back online.   &lt;br /&gt;
&lt;br /&gt;
'''SciNet Downtime July 20th, 2021 (Tuesday):''' There will be a maintenance shutdown of the SciNet data center on Tuesday July 20th, starting at 7 am EDT. There will be no access to any of the SciNet systems (Niagara, Mist, HPSS, Teach cluster, or the file systems) during this time.  We expect to be able to bring the systems back online in the evening of July 20th.  The status of the Niagara cluster can be checked on status.computecanada.ca. For up-to-date and more detailed information on the status of all the SciNet systems, you can always check back here.&lt;br /&gt;
&lt;br /&gt;
'''June 28th, 2021, 4:06 PM:''' Mist OS upgrade is complete.&lt;br /&gt;
&lt;br /&gt;
'''May 27, 2021:''' Datamovers addresses have changed to improve high bandwidth connectivity and cybersecurity. The new addresses are 142.1.174.227 for nia-datamover1.scinet.utoronto.ca, and 142.1.174.228 for nia-datamover2.scinet.utoronto.ca.&lt;br /&gt;
&lt;br /&gt;
If you have jobs that need to connect to a software license server using an ssh tunnel through nia-gw (which actually resolves to datamover1 or datamover2), you may need to ask the system administrators of that license server to allow incoming connections from the new addresses above.&lt;br /&gt;
'''June 29th, 2021, 2:00 PM:''' Thunderstorm-related power fluctuations are causing some Niagara compute nodes and their jobs to crash.  Please resubmit if your jobs seem to have crashed for no apparent reason.&lt;br /&gt;
&lt;br /&gt;
'''June 28th, 2021, 4:06 PM:''' Mist OS upgrade is complete.&lt;br /&gt;
&lt;br /&gt;
'''June 28th, 2021, 9:00 AM:''' Mist is under maintenance. OS upgrading from RHEL 7 to 8.&lt;br /&gt;
&lt;br /&gt;
'''June 11th, 2021, 8:30 AM:''' Maintenance complete. Systems are up.&lt;br /&gt;
&lt;br /&gt;
'''June 9th to 10th, 2021:''' The SciNet datacentre will have a scheduled maintenance shutdown.  Niagara, Mist, Rouge, HPSS, login nodes, the file systems, and hosted systems will all be offline during the shutdown starting at 7AM EDT on Wednesday June 9th. We expect the systems to be back up in the morning of Friday June 11th.  Check here for updates.&lt;br /&gt;
&lt;br /&gt;
'''May 27, 2021:''' Datamovers addresses have changed to improve high bandwidth connectivity and cybersecurity. The new addresses are 142.1.174.227 for nia-datamover1.scinet.utoronto.ca, and 142.1.174.228 for nia-datamover2.scinet.utoronto.ca.&lt;br /&gt;
&lt;br /&gt;
'''May 27th, 20:00.''' All systems are up and running &lt;br /&gt;
&lt;br /&gt;
'''May 27th, 19:30.''' Most systems are up&lt;br /&gt;
&lt;br /&gt;
'''May 27th, 19:00:''' Cooling is back. Powering up systems&lt;br /&gt;
&lt;br /&gt;
'''May 27th, 2021, 11:30am:'''  The cooling tower issue has been identified as a wiring issue and is being repaired.  We don't have an ETA on when cooling will be restored, however we are hopeful it will be by the end of the day.  &lt;br /&gt;
&lt;br /&gt;
'''May 27th, 2021, 12:30am:''' Cooling tower motor is not working properly and may need to be replaced.  Its the primary motor and the cooling system can not run without it, so at least until tomorrow all equipment at the datacenter will remain unavailable.  Updates about expected repair times will be posted when they are known.&lt;br /&gt;
&lt;br /&gt;
'''May 26th, 2021, 9:20pm:''' we are currently experiencing cooling issues at the SciNet data centre.  Updates will be posted as we determine the cause of the problem.&lt;br /&gt;
&lt;br /&gt;
'''From Tue Mar 30 at 12 noon EST to Thu Apr 1 at 12 noon EST,''' there will be a two-day reservation for the &amp;quot;Niagara at Scale&amp;quot; pilot  event.  During these 48 hours, only &amp;quot;Niagara at Scale&amp;quot; projects will run on the compute notes (as well as SOSCIP projects, on a subset of nodes).  All other users can still login, access their data, and submit jobs throughout this event, but the jobs will not run until after the event.  The debugjob queue will remain available to  everyone as well.&lt;br /&gt;
&lt;br /&gt;
The scheduler will not start batch jobs that cannot finish before the start of this event. Users can submit small and short jobs can take advantage of this, as the scheduler may be able to fit these jobs in before the event starts on the otherwise idle nodes.&lt;br /&gt;
&lt;br /&gt;
'''Tue 23 Mar 2021 12:19:07 PM EDT''' - Planned external network maintenance 12pm-1pm Tuesday, March 23rd. &lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 28 17:35:16 EST 2021:''' HPSS services are back online&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 28 12:36:21 EST 2021:''' HPSS services offline&lt;br /&gt;
&lt;br /&gt;
We need a small maintenance window as early as possible still this afternoon to perform a small change in configuration. Ongoing jobs will be allowed to finish, but we are keeping new submissions on hold on the queue.&lt;br /&gt;
&lt;br /&gt;
'''Mon Jan 25 13:16:33 EST 2021:''' HPSS services are back online&lt;br /&gt;
&lt;br /&gt;
'''Sat Jan 23 10:03:33 EST 2021:''' HPSS services offline&lt;br /&gt;
&lt;br /&gt;
We detected some type of hardware failure on our HPSS equipment overnight, so access has been disabled pending further investigation.&lt;br /&gt;
&lt;br /&gt;
'''Fri Jan 22 10:49:29 EST 2021:''' The Globus transition to oauth is finished&lt;br /&gt;
&lt;br /&gt;
Please deactivate any previous sessions to the niagara endpoint (in the last 7 days), and activate/login again. &lt;br /&gt;
&lt;br /&gt;
For more details check https://docs.scinet.utoronto.ca/index.php/Globus#computecandada.23niagara&lt;br /&gt;
&lt;br /&gt;
'''Jan 21, 2021:''' Globus access disruption on Fri, Jan/22/2021 10AM: Please be advised that we will have a maintenance window starting tomorrow at 10AM to roll out the transition of services to oauth based authentication.&lt;br /&gt;
&lt;br /&gt;
'''Jan 15, 2021:'''Globus access update on Mon, Jan/18/2021 and Tue, Jan/19/2021:&lt;br /&gt;
Please be advised we start preparations on Monday to perform update to Globus access on Tuesday. We'll be adopting oauth instead of myproxy from that point on. During this period expect sporadic disruptions of service. On Monday we'll already block access to nia-dm2, so please refrain from starting new login sessions or ssh tunnels via nia-dm2 from this weekend already.&lt;br /&gt;
&lt;br /&gt;
''' December 11,2020, 12:00 AM EST: ''' Cooling issue resolved. Systems back.&lt;br /&gt;
&lt;br /&gt;
''' December 11,2020, 6:00 PM EST: ''' Cooling issue at datacenter. All systems down.&lt;br /&gt;
&lt;br /&gt;
''' December 7, 2020, 7:25 PM EST: '''All systems back; users can log in again.&lt;br /&gt;
&lt;br /&gt;
''' December 7, 2020, 6:46 PM EST: '''User connectivity to data center not yet ready, but queued jobs on Mist and Niagara have been started.&lt;br /&gt;
 &lt;br /&gt;
''' December 7, 2020, 7:00 AM EST: '''Maintenance shutdown in effect. This is a one-day maintenance shutdown.  There will be no access to Niagara, Mist, HPSS or teach, nor to their file systems during this time.  We expect to be able to bring the systems back online this evening.&lt;br /&gt;
&lt;br /&gt;
''' December 2, 2020, 9:10 PM EST: '''Power is back, systems are coming up. Please resubmit any jobs that failed because of this incident.&lt;br /&gt;
&lt;br /&gt;
''' December 2, 2020, 6:00 PM EST: '''Power glitch at the data center, caused about half of the compute nodes to go down.  Power issue not yet resolved.&lt;br /&gt;
&lt;br /&gt;
'''&amp;lt;span style=&amp;quot;color:#dd1111&amp;quot;&amp;gt;Announcing a Maintenance Shutdown on December 7th, 2020&amp;lt;/span&amp;gt;''' &amp;lt;br/&amp;gt;There will be a one-day maintenance shutdown on December 7th 2020, starting at 7 am EST.  There will be no access to Niagara, Mist, HPSS or teach, nor to their file systems during this time.  We expect to be able to bring the systems back online in the evening of the same day.&lt;br /&gt;
&lt;br /&gt;
''' November 6, 2020, 8:00 PM EST: ''' Systems are coming back online.&lt;br /&gt;
&lt;br /&gt;
''' November 6, 2020, 9:49 AM EST: ''' Repairs on the cooling system are underway.  No ETA, but the systems will likely be back some time today.&lt;br /&gt;
&lt;br /&gt;
''' November 6, 2020, 4:27 AM EST: '''Cooling system failure, datacentre is shut down.&lt;br /&gt;
&lt;br /&gt;
''' October 9, 2020, 12:57 PM: ''' A short power glitch caused many of the Niagara compute nodes to lose power; jobs running on them would have failed. Please check your jobs and resubmit.&lt;br /&gt;
&lt;br /&gt;
''' October 8, 2020, 9:50 PM: ''' Jupyterhub service is back up.&lt;br /&gt;
&lt;br /&gt;
''' October 8, 2020, 5:40 PM: ''' Jupyterhub service is down. We are investigating.&lt;br /&gt;
&lt;br /&gt;
''' September 28, 2020, 11:00 AM EST: ''' A short power glitch caused many of the Niagara compute nodes to lose power; jobs running on them would have failed. Please check your jobs and resubmit.&lt;br /&gt;
&lt;br /&gt;
''' September 1, 2020, 2:15 PM EST: ''' A short power glitch caused about half of the Niagara compute nodes to lose power; jobs running on them would have failed. Please check your jobs and resubmit.&lt;br /&gt;
&lt;br /&gt;
''' September 1, 2020, 9:27 AM EST: ''' The Niagara cluster has moved to a new default software stack, NiaEnv/2019b.  If your job scripts used the previous default software stack before (NiaEnv/2018a), please put the command &amp;quot;module load NiaEnv/2018a&amp;quot; before other module commands in those scripts, to ensure they will continue to work, or try the new stack (recommended).&lt;br /&gt;
''' August 24, 2020, 7:37 PM EST: ''' Connectivity is back to normal&lt;br /&gt;
&lt;br /&gt;
''' August 24, 2020, 6:35 PM EST: ''' We have partial connectivity back, but are still investigating.&lt;br /&gt;
&lt;br /&gt;
''' August 24, 2020, 3:15 PM EST: ''' There are issues connecting to the data centre. We're investigating.&lt;br /&gt;
&lt;br /&gt;
''' August 21, 2020, 6:00 PM EST: ''' The pump has been repaired, cooling is restored, systems are up.  &amp;lt;br/&amp;gt;Scratch purging is postponed until the evening of Friday Aug 28th, 2020.&lt;br /&gt;
&lt;br /&gt;
'''August 19, 2020, 4:40 PM EST:''' Update: The current estimate is to have the cooling restored on Friday and we hope to have the systems available for users on Saturday August 22, 2020.&lt;br /&gt;
&lt;br /&gt;
'''August 17, 2020, 4:00 PM EST:''' Unfortunately after taking the pump apart it was determined there was a more serious failure of the main drive shaft, not just the seal. As a new one will need to be sourced or fabricated we're estimating that it will take at least a few more days to get the part and repairs done to restore cooling. Sorry for the inconvenience. &lt;br /&gt;
&lt;br /&gt;
'''August 15, 2020, 1:00 PM EST:''' Due to parts availablity to repair the failed pump and cooling system it is unlikely that systems will be able to be restored until Monday afternoon at the earliest. &lt;br /&gt;
&lt;br /&gt;
'''August 15, 2020, 00:04 AM EST:'''  A primary pump seal in the cooling infrastructure has blown and parts availability will not be able be determined until tomorrow. All systems are shut down as there is no cooling.  If parts are available, systems may be back at the earliest late tomorrow. Check here for updates.  &lt;br /&gt;
&lt;br /&gt;
'''August 14, 2020, 21:04 AM EST:''' Tomorrow's /scratch purge has been postponed.&lt;br /&gt;
&lt;br /&gt;
'''August 14, 2020, 21:00 AM EST:''' Staff at the datacenter. Looks like one of the pumps has a seal that is leaking badly.&lt;br /&gt;
&lt;br /&gt;
'''August 14, 2020, 20:37 AM EST:''' We seem to be undergoing a thermal shutdown at the datacenter.&lt;br /&gt;
&lt;br /&gt;
'''August 14, 2020, 20:20 AM EST:''' Network problems to niagara/mist. We are investigating.&lt;br /&gt;
 &lt;br /&gt;
'''August 13, 2020, 10:40 AM EST:''' Network is fixed, scheduler and other services are back.&lt;br /&gt;
&lt;br /&gt;
'''August 13, 2020, 8:20 AM EST:''' We had an IB switch failure, which is affecting a subset of nodes, including the scheduler nodes.&lt;br /&gt;
&lt;br /&gt;
'''August 10, 2020, 7:30 PM EST:''' Scheduler fully operational again.&lt;br /&gt;
&lt;br /&gt;
'''August 10, 2020, 3:00 PM EST:''' Scheduler partially functional: jobs can be submitted and are running.&lt;br /&gt;
&lt;br /&gt;
'''August 10, 2020, 2:00 PM EST:''' Scheduler is temporarily inoperational.&lt;br /&gt;
&lt;br /&gt;
'''August 7, 2020, 9:15 PM EST:''' Network is fixed, scheduler and other services are coming back.&lt;br /&gt;
&lt;br /&gt;
'''August 7, 2020, 8:20 PM EST:''' Disruption of part of the network in the data centre.  Causes issue with the scheduler, the mist login node, and possibly others. We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''July 30, 2020, 9:00 AM''' Project backup in progress but incomplete: please be aware that after we deployed the new, larger storage appliance for scratch and project two months ago, we started a full backup of project (1.5PB). This backup is taking a while to complete, and there are still a few areas which have not been backed up fully. Please be careful to not delete things from project that you still need, in particular if they are recently added material.&lt;br /&gt;
&lt;br /&gt;
'''July 27, 2020, 5:00 PM:''' Scheduler issues resolved.&lt;br /&gt;
&lt;br /&gt;
'''July 27, 2020, 3:00 PM:''' Scheduler issues. We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''July 13, 4:40 PM:''' Most systems are available again. Only Mist is still being brought up.&lt;br /&gt;
&lt;br /&gt;
'''July 13, 10:00 AM:''' '''SciNet/Niagara Downtime In Progress'''&lt;br /&gt;
&lt;br /&gt;
'''SciNet/Niagara Downtime Announcement, July 13, 2020'''&amp;lt;br/&amp;gt;&lt;br /&gt;
All resources at SciNet will undergo a maintenance shutdown on Monday July 13, 2020, starting at 10:00 am EDT, for file system and scheduler upgrades.  There will be no access to any of the SciNet systems (Niagara, Mist, HPSS, Teach cluster, or the file systems) during this time.&lt;br /&gt;
We expect to be able to bring the systems back around 3 PM (EST) on the same day.&lt;br /&gt;
&lt;br /&gt;
''' June 29, 6:21:00  PM:''' Systems are available again.  &lt;br /&gt;
&lt;br /&gt;
''' June 29, 12:30:00  PM:''' Power Outage caused thermal shutdown.&lt;br /&gt;
&lt;br /&gt;
'''June 20, 2020, 10:24 PM:''' File systems are back up.  Unfortunately, all running jobs would have died and users are asked to resubmit them.&lt;br /&gt;
&lt;br /&gt;
'''June 20, 2020, 9:48 PM:''' An issue with the file systems is causing trouble.  We are investigating the cause.&lt;br /&gt;
&lt;br /&gt;
'''June 15, 2020, 10:30 PM:''' A '''power glitch''' caused some compute nodes to be rebooted: jobs running at the time may have failed; users are asked to resubmit these jobs.&lt;br /&gt;
&lt;br /&gt;
'''June 12, 2020, 6:15 PM:''' Two '''power glitches''' during the night caused some compute nodes to be rebooted: jobs running at the time may have failed; users are asked to resubmit these jobs.&lt;br /&gt;
&lt;br /&gt;
'''June 6, 2020, 6:06 AM:''' A '''power glitch''' caused some compute nodes to be rebooted: jobs running at the time may have failed; users are asked to resubmit these jobs.&lt;br /&gt;
&lt;br /&gt;
'''May 24, 2020, 8:20 AM:''' A '''power glitch''' this morning caused all compute nodes to be rebooted: jobs running at the time may have failed; users are asked to resubmit these jobs.&lt;br /&gt;
&lt;br /&gt;
'''May 7, 2020, 6:05 PM:''' Maintenance shutdown is finished.  Most systems are back in production.&lt;br /&gt;
&lt;br /&gt;
'''May 6, 2020, 7:08 AM:''' Two-day datacentre maintenance shutdown has started.&lt;br /&gt;
&lt;br /&gt;
''' SciNet/Niagara Downtime Announcement, May 6-7, 2020'''&lt;br /&gt;
&lt;br /&gt;
All resources at SciNet will undergo a two-day maintenance shutdown on May 6th and 7th 2020, starting at 7 am EDT on Wednesday May 6th.  There will be no access to any of the SciNet systems (Niagara, Mist, HPSS, Teach cluster, or the file systems) or systems hosted at the SciNet data centre.  We expect to be able to bring the systems back online the evening of May 7th.&lt;br /&gt;
&lt;br /&gt;
'''May 4, 2020, 7:51 AM:''' A power glitch this morning caused compute nodes to be rebooted: jobs running at the time may have failed; users are asked to resubmit these jobs.&lt;br /&gt;
&lt;br /&gt;
'''May 3, 2020, 8:20 AM:''' A power glitch this morning caused all compute nodes to be rebooted: jobs running at the time may have failed; users are asked to resubmit these jobs.&lt;br /&gt;
&lt;br /&gt;
'''April 28, 2020, 7:20 AM:''' A power glitch this morning caused all compute nodes to be rebooted: jobs running at the time have failed; users are asked to resubmit these jobs.&lt;br /&gt;
 &lt;br /&gt;
'''April 20, 2020: Security Incident at Cedar; implications for Niagara users'''&lt;br /&gt;
&lt;br /&gt;
Last week, it became evident that the Cedar GP cluster had been&lt;br /&gt;
comprimised for several weeks.  The passwords of at least two&lt;br /&gt;
Compute Canada users were known to the attackers. One of these was&lt;br /&gt;
used to escalate privileges on Cedar, as explained on&lt;br /&gt;
https://status.computecanada.ca/view_incident?incident=423.&lt;br /&gt;
&lt;br /&gt;
These accounts were used to login to Niagara as well, but Niagara&lt;br /&gt;
did not have the same security loophole as Cedar (which has been&lt;br /&gt;
fixed), and no further escalation was observed on Niagara.&lt;br /&gt;
&lt;br /&gt;
Reassuring as that may sound, it is not known how the passwords of&lt;br /&gt;
the two user accounts were obtained. Given this uncertainty, the&lt;br /&gt;
SciNet team *strongly* recommends that you change your password on&lt;br /&gt;
https://ccdb.computecanada.ca/security/change_password, and remove&lt;br /&gt;
any SSH keys and regenerate new ones (see&lt;br /&gt;
https://docs.scinet.utoronto.ca/index.php/SSH_keys).&lt;br /&gt;
&lt;br /&gt;
''' Tue 30 Mar 2020 14:55:14 EDT'''  Burst Buffer available again.&lt;br /&gt;
&lt;br /&gt;
''' Fri Mar 27 15:29:00 EDT 2020:''' SciNet systems are back up. Only the Burst Buffer remains offline, its maintenance is expected to be finished early next week.&lt;br /&gt;
&lt;br /&gt;
''' Thu Mar 26 23:05:00 EDT 2020:'''  Some aspects of the maintenance took longer than expected. The systems will not be back up until some time tomorrow, Friday March 27, 2020.  &lt;br /&gt;
&lt;br /&gt;
''' Wed Mar 25 7:00:00 EDT 2020:'''  SciNet/Niagara downtime started.&lt;br /&gt;
&lt;br /&gt;
''' Mon Mar 23 18:45:10 EDT 2020:'''  File system issues were resolved.&lt;br /&gt;
&lt;br /&gt;
''' Mon Mar 23 18:01:19 EDT 2020:''' There is currently an issue with the main Niagara filesystems. This effects all systems, all jobs have been killed. The issue is being investigated. &lt;br /&gt;
&lt;br /&gt;
''' Fri Mar 20 13:15:33 EDT 2020: ''' There was a power glitch at the datacentre at 8:50 AM, which resulted in jobs getting killed.  Please resubmit failed jobs. &lt;br /&gt;
&lt;br /&gt;
''' COVID-19 Impact on SciNet Operations, March 18, 2020'''&lt;br /&gt;
&lt;br /&gt;
Although the University of Toronto is closing of some of its&lt;br /&gt;
research operations on Friday March 20 at 5 pm EDT, this does not&lt;br /&gt;
affect the SciNet systems (such as Niagara, Mist, and HPSS), which&lt;br /&gt;
will remain operational.&lt;br /&gt;
&lt;br /&gt;
''' SciNet/Niagara Downtime Announcement, March 25-26, 2020'''&lt;br /&gt;
&lt;br /&gt;
All resources at SciNet will undergo a two-day maintenance shutdown on March 25th and 26th 2020, starting at 7 am EDT on Wednesday March 25th.  There will be no access to any of the SciNet systems (Niagara, Mist, HPSS, Teach cluster, or the file systems) during this time.&lt;br /&gt;
&lt;br /&gt;
This shutdown is necessary to finish the expansion of the Niagara cluster and its storage system.&lt;br /&gt;
&lt;br /&gt;
We expect to be able to bring the systems back online the evening of March 26th.&lt;br /&gt;
&lt;br /&gt;
''' March 9, 2020, 11:24 PM:''' HPSS services are temporarily suspended for emergency maintenance.&lt;br /&gt;
&lt;br /&gt;
''' March 7, 2020, 10:15 PM:''' File system issues have been cleared.&lt;br /&gt;
&lt;br /&gt;
''' March 6, 2020, 7:30 PM:''' File system issues; we are investigating&lt;br /&gt;
&lt;br /&gt;
''' March 2, 2020, 1:30 PM:''' For the extension of Niagara, the operating system on all Niagara nodes has been upgraded&lt;br /&gt;
from CentOS 7.4 to 7.6.  This required all&lt;br /&gt;
nodes to be rebooted. Running compute jobs are allowed to finish&lt;br /&gt;
before the compute node gets rebooted. Login nodes have all been rebooted, as have the datamover nodes and the jupyterhub service.&lt;br /&gt;
&lt;br /&gt;
''' Feb 24, 2020, 1:30PM: ''' The [[Mist]] login node got rebooted.  It is back, but we are still monitoring the situation.&lt;br /&gt;
&lt;br /&gt;
''' Feb 12, 2020, 11:00AM: ''' The [[Mist]] GPU cluster now available to users.&lt;br /&gt;
&lt;br /&gt;
''' Feb 11, 2020, 2:00PM: ''' The Niagara compute nodes were accidentally rebooted, killing all running jobs.&lt;br /&gt;
&lt;br /&gt;
''' Feb 10, 2020, 19:00PM: ''' HPSS is back to normal.&lt;br /&gt;
&lt;br /&gt;
''' Jan 30, 2020, 12:01PM: ''' We are having an issue with HPSS, in which the disk-cache is full. We put a reservation on the whole system (Globus, plus archive and vfs queues), until it has had a chance to clear some space on the cache.&lt;br /&gt;
&lt;br /&gt;
''' Jan 21, 2020, 4:05PM: '''   The was a partial power outage the took down a large amount of the compute nodes.  If your job died during this period please resubmit.  &lt;br /&gt;
&lt;br /&gt;
'''Jan 13, 2020, 7:35 PM:''' Maintenance finished.&lt;br /&gt;
&lt;br /&gt;
'''Jan 13, 2020, 8:20 AM:''' The announced maintenance downtime started (see below).&lt;br /&gt;
&lt;br /&gt;
'''Jan 9 2020, 11:30 AM:''' External ssh connectivity restored, issue related to the university network.&lt;br /&gt;
&lt;br /&gt;
'''Jan 9 2020, 9:24 AM:''' We received reports of users having trouble connecting into the SciNet data centre; we're investigating.  Systems are up and running and jobs are fine.&amp;lt;p&amp;gt;&lt;br /&gt;
As a work around, in the meantime, it appears to be possible to log into graham, cedar or beluga, and then ssh to niagara.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Downtime announcement:'''&lt;br /&gt;
To prepare for the upcoming expansion of Niagara, there will be a&lt;br /&gt;
one-day maintenance shutdown on '''January 13th 2020, starting at 8 am&lt;br /&gt;
EST'''.  There will be no access to Niagara, Mist, HPSS or teach, nor&lt;br /&gt;
to their file systems during this time.&lt;br /&gt;
&lt;br /&gt;
2019&lt;br /&gt;
&lt;br /&gt;
'''December 13, 9:00 AM EST:''' Issues resolved.&lt;br /&gt;
&lt;br /&gt;
'''December 13, 8:20 AM EST:''' Overnight issue is now preventing logins to Niagara and other services. Possibly a file system issue, we are investigating.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt; '''Fri, Nov 15 2019, 11:00 PM (EST)'''  Niagara and most of the main systems are now available. &lt;br /&gt;
&amp;lt;/p&amp;gt;&amp;lt;p&amp;gt; '''Fri, Nov 15 2019, 7:50 PM (EST)'''  SOSCIP GPU cluster is up and accessible.  Work on the other systems continues.&lt;br /&gt;
&amp;lt;/p&amp;gt;&amp;lt;p&amp;gt; '''Fri, Nov 15 2019, 5:00 PM (EST)'''  Infrastructure maintenance done, upgrades still in process.&lt;br /&gt;
&amp;lt;/p&amp;gt;&amp;lt;p&amp;gt;&lt;br /&gt;
'''Fri, Nov 15 2019, 7:00 AM (EST)'''  Maintenance shutdown of the SciNet data centre has started.  Note: scratch purging has been postponed until Nov 17.&amp;lt;br/&amp;gt; &lt;br /&gt;
&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&lt;br /&gt;
'''Announcement:''' &lt;br /&gt;
The SciNet datacentre will undergo a maintenance shutdown on&lt;br /&gt;
Friday November 15th 2019, from 7 am to 11 pm (EST), with no access&lt;br /&gt;
to any of the SciNet systems (Niagara, P8, SGC, HPSS, Teach cluster,&lt;br /&gt;
or the filesystems) during that time. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
'''Sat, Nov 2 2019, 1:30 PM (update):'''  Chiller has been fixed, all systems are operational.    &lt;br /&gt;
&amp;lt;/p&amp;gt;&lt;br /&gt;
'''Fri, Nov 1 2019, 4:30 PM (update):'''  We are operating in free cooling so have brought up about 1/2 of the Niagara compute nodes to reduce the cooling load.  Access, storage, and other systems should now be available.   &lt;br /&gt;
&lt;br /&gt;
'''Fri, Nov 1 2019, 12:05 PM (update):''' A power module in the chiller has failed and needs to be replaced.   We should be able to operate in free cooling if the temperature stays cold enough, but we may not be able to run all systems. No eta yet on when users will be able to log back in. &lt;br /&gt;
&lt;br /&gt;
'''Fri, Nov 1 2019, 9:15 AM (update):''' There was a automated shutdown because of rising temperatures, causing all systems to go down. We are investigating, check here for updates.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;'''Fri, Nov 1 2019, 8:16 AM:''' Unexpected data centre issue: Check here for updates.&lt;br /&gt;
&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
''' Thu 1 Aug 2019 5:00:00 PM ''' Systems are up and operational.   &lt;br /&gt;
&lt;br /&gt;
'''Thu 1 Aug 2019 7:00:00 AM: ''' Scheduled Downtime Maintenance of the SciNet Datacenter.  All systems will be down and unavailable starting 7am until the evening. &lt;br /&gt;
&lt;br /&gt;
'''Fri 26 Jul 2019, 16:02:26 EDT:''' There was an issue with the Burst Buffer at around 3PM, and it was recently solved. BB is OK again.&lt;br /&gt;
&lt;br /&gt;
''' Sun 30 Jun 2019 ''' The '''SOSCIP BGQ''' and '''P7''' systems were decommissioned on '''June 30th, 2019'''.  The BGQdev front end node and storage are still available.  &lt;br /&gt;
&lt;br /&gt;
'''Wed 19 Jun 2018, 1:20:00 PM:''' The BGQ is back online.&lt;br /&gt;
&lt;br /&gt;
'''Wed 19 Jun 2018, 10:00:00 AM:''' The BGQ is still down, the SOSCIP GPU nodes should be back up. &lt;br /&gt;
&lt;br /&gt;
'''Wed 19 Jun 2018, 1:40:00 AM:''' There was an issue with the SOSCIP BGQ and GPU Cluster last night about 1:42am, probably a power fluctuation that took it down.  &lt;br /&gt;
&lt;br /&gt;
'''Wed 12 Jun 2019, 3:30 AM - 7:40 AM''' Intermittent system issues on Niagara's project and scratch as the file number limit was reached. We increased the number of files allowed in total on the file system. &lt;br /&gt;
&lt;br /&gt;
'''Thu 30 May 2019, 11:00:00 PM:'''&lt;br /&gt;
The maintenance downtime of SciNet's data center has finished, and systems are being brought online now.  You can check the progress here. Some systems might not be available until Friday morning.&amp;lt;br/&amp;gt;&lt;br /&gt;
Some action on the part of users will be required when they first connect again to a Niagara login nodes or datamovers.  This is due to the security upgrade of the Niagara cluster, which is now in line with currently accepted best practices.&amp;lt;br/&amp;gt;&lt;br /&gt;
The details of the required actions can be found on the [[SSH Changes in May 2019]] wiki page.&lt;br /&gt;
&lt;br /&gt;
'''Wed 29-30 May 2019''' The SciNet datacentre will undergo a two-day maintenance shutdown, starting at 7 am EDT on Wednesday May 29th.  There will be no access to any of the SciNet systems (Niagara, P7, P8, BGQ, SGC, HPSS, Teach cluster, or the file systems) during this time.&lt;br /&gt;
&lt;br /&gt;
'''SCHEDULED SHUTDOWN''': &lt;br /&gt;
&lt;br /&gt;
Please be advised that on '''Wednesday May 29th through Thursday May 30th''', the SciNet datacentre will undergo a two-day maintenance shutdown, starting at 7 am EDT on Wednesday May 29th.  There will be no access to any of the SciNet systems (Niagara, P7, P8, BGQ, SGC, HPSS, Teach cluster, or the file systems) during this time.&lt;br /&gt;
&lt;br /&gt;
This is necessary to finish the installation of an emergency power generator, to perform the annual cooling tower maintenance, and to enhance login security.&lt;br /&gt;
&lt;br /&gt;
We expect to be able to bring the systems back online the evening of May 30th.  Due to the enhanced login security, the ssh applications of users will need to update their known host list. More detailed information on this procedure will be sent shortly before the systems are back online.&lt;br /&gt;
&lt;br /&gt;
'''Fri 5 Apr 2019:''' Software updates on Niagara: The default CCEnv software stack now uses avx512 on Niagara, and there is now a NiaEnv/2019b stack (&amp;quot;epoch&amp;quot;). &lt;br /&gt;
&lt;br /&gt;
'''Thu 4 Apr 2019:''' The 2019 compute and storage allocations have taken effect on Niagara.&lt;br /&gt;
&lt;br /&gt;
'''NOTE''':  There is scheduled network maintenance for '''Friday April 26th 12am-8am''' on the Scinet datacenter external network connection.   This will not affect internal connections and running jobs however remote connections may see interruptions during this period.&lt;br /&gt;
&lt;br /&gt;
'''Wed 24 Apr 2019 14:14 EDT:''' HPSS is back on service. Library and robot arm maintenance finished.&lt;br /&gt;
&lt;br /&gt;
'''Wed 24 Apr 2019 08:35 EDT:''' HPSS out of service this morning for library and robot arm maintenance.&lt;br /&gt;
&lt;br /&gt;
'''Fri 19 Apr 2019 17:40 EDT:''' HPSS robot arm has been released and is back to normal operations.&lt;br /&gt;
&lt;br /&gt;
'''Fri 19 Apr 2019 14:00 EDT:''' problems with HPPS library robot have been detected.&lt;br /&gt;
&lt;br /&gt;
'''Wed 17 Apr 2019 15:35 EDT:''' Network connection is back.&lt;br /&gt;
&lt;br /&gt;
'''Wed 17 Apr 2019 15:12 EDT:''' Network connection down.  Investigating.&lt;br /&gt;
&lt;br /&gt;
'''Tue 9 Apr 2019 22:24:14 EDT:'''  Network connection restored.&lt;br /&gt;
&lt;br /&gt;
'''Tue 9 Apr 2019, 15:20:''' Network connection down.  Investigating.&lt;br /&gt;
&lt;br /&gt;
'''Fri 5 Apr 2019:''' Planned, short outage in connectivity to the SciNet datacentre from 7:30 am to 8:55 am EST for maintenance of the network.  This outage will not affect running or queued jobs. It may be necessary to reboot the login nodes at some point tomorrow, which could result in a short interruption of connectivity, but which will have no effect on running or queued jobs.&lt;br /&gt;
&lt;br /&gt;
'''April 4, 2019:'''  The 2019 compute and storage allocations will take effect on Niagara. Running jobs will not be affected by this change and will run their course.  Queued jobs' priorities will be updated to reflect the new fairshare values later in the day.  The queue should fully reflect the new fairshare values in about 24 hours.   &lt;br /&gt;
&lt;br /&gt;
It may be necessary to reboot the login nodes at some point tomorrow, which could result in a short interruption of connectivity, but which will have no effect on running or queued jobs.&lt;br /&gt;
&lt;br /&gt;
There will be updates to the software stack on this day as well.&lt;br /&gt;
&lt;br /&gt;
'''March 25, 3:05 PM EST:'''  Most systems back online, other services should be back shortly. &lt;br /&gt;
&lt;br /&gt;
'''March 25, 12:05 PM EST:''' Power is back at the datacentre, but it is not yet known when all systems will be back up.  Keep checking here for updates.&lt;br /&gt;
&lt;br /&gt;
'''March 25, 11:27 AM EST:''' A power outage in the datacentre occured and caused all services to go down.  Check here for updates.&lt;br /&gt;
&lt;br /&gt;
'''Thu Mar 21 10:37:28 EDT 2019:''' HPSS is back in service&lt;br /&gt;
&lt;br /&gt;
HPSS out of service on '''Tue, Mar/19 at 9AM''', for tape library expansion and relocation. It's possible the downtime will extend to Wed, Mar/20.&lt;br /&gt;
&lt;br /&gt;
'''January 21, 4:00 PM''': HPSS is back in service. Thank you for your patience.&lt;br /&gt;
&lt;br /&gt;
'''January 18, 5:00 PM''': We did practically all of the HPSS upgrades (software/hardware), however the main client node - archive02 - is presenting an issue we just couldn't resolve yet. We will try to resume work over the weekend with cool heads, or on Monday. Sorry, but this is an unforeseen delay. Jobs on the queue we'll remain there, and we'll delay the scratch purging by 1 week.&lt;br /&gt;
&lt;br /&gt;
'''January 16, 11:00 PM''': HPSS is being upgraded, as announced.&lt;br /&gt;
&lt;br /&gt;
'''January 16, 8:00 PM''': System are coming back up and should be accessible for users now.&lt;br /&gt;
&lt;br /&gt;
'''January 15, 8:00 AM''': Data centre downtime in effect.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;font color=red&amp;gt;&amp;lt;b&amp;gt;Downtime Announcement for January 15 and 16, 2019&amp;lt;/b&amp;gt;&amp;lt;/font&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
The SciNet datacentre will need to undergo a two-day maintenance shutdown in order to perform electrical work, repairs and maintenance.  The electrical work is in preparation for the upcoming installation of an emergency power generator and a larger UPS, which will result in increased resilience to power glitches and outages.  The shutdown is scheduled to start on '''Tuesday January 15, 2019, at 7 am''' and will last until '''Wednesday 16, 2019''', some time in the evening. There will be no access to any of the SciNet systems (Niagara, P7, P8, BGQ, SGC, HPSS, Teach cluster, or the filesystems) during this time.&lt;br /&gt;
Check back here for up-to-date information on the status of the systems.&lt;br /&gt;
&lt;br /&gt;
Note: this downtime was originally scheduled for Dec. 18, 2018, but has been postponed and combined with the annual maintenance downtime.&lt;br /&gt;
&lt;br /&gt;
'''December 24, 2018, 11:35 AM EST:''' Most systems are operational again. If you had compute jobs running yesterday at around 3:30PM, they likely crashed - please check them and resubmit if needed.&lt;br /&gt;
&lt;br /&gt;
'''December 24, 2018, 10:40 AM EST:''' Repairs have been made, and the file systems are starting to be mounted on the cluster. &lt;br /&gt;
&lt;br /&gt;
'''December 23, 2018, 3:38 PM EST:''' Issues with the file systems (home, scratch and project). We are investigating, it looks like a hardware issue that we are trying to work around. Note that the absence of /home means you cannot log in with ssh keys. All compute jobs crashed around 3:30 PM EST on Dec 23. Once the system is properly up again, please resubmit your jobs.  Unfortunately, at this time of year, it is not possible to give an estimate on when the system will be operational again.&lt;br /&gt;
&lt;br /&gt;
'''Tue Nov 22 14:20:00 EDT 2018''': &amp;lt;font color=green&amp;gt;HPSS back in service&amp;lt;/font&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Tue Nov 22 08:55:00 EDT 2018''': &amp;lt;font color=red&amp;gt;HPSS offline for scheduled maintenance&amp;lt;/font&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Tue Nov 20 16:30:00 EDT 2018''':  HPSS offline on Thursday 9AM for installation of new LTO8 drives in the tape library.&lt;br /&gt;
&lt;br /&gt;
'''Tue Oct  9 12:16:00 EDT 2018''':  BGQ compute nodes are up.  &lt;br /&gt;
&lt;br /&gt;
'''Sun Oct  7 20:24:26 EDT 2018''':  SGC and BGQ front end are available,  BGQ compute nodes down related to a cooling issue.  &lt;br /&gt;
&lt;br /&gt;
'''Sat Oct  6 23:16:44 EDT 2018''':  There were some problems bringing up SGC &amp;amp; BGQ, they will remain offline for now.&lt;br /&gt;
&lt;br /&gt;
'''Sat Oct  6 18:36:35 EDT 2018''':  Electrical work finished, power restored. Systems are coming online.&lt;br /&gt;
&lt;br /&gt;
'''July 18, 2018:''' login.scinet.utoronto.ca is now disabled, GPC $SCRATCH and $HOME are decommissioned.&lt;br /&gt;
&lt;br /&gt;
'''July 12, 2018:''' There was a short power interruption around 10:30 am which caused most of the systems (Niagara, SGC, BGQ) to reboot and any running jobs to fail. &lt;br /&gt;
&lt;br /&gt;
'''July 11, 2018:''' P7's moved to BGQ filesystem, P8's moved to Niagara filesystem.&lt;br /&gt;
&lt;br /&gt;
'''May 24, 2018, 9:25 PM EST:''' The data center is up, and all systems are operational again.&lt;br /&gt;
&lt;br /&gt;
'''May 24, 2018, 7:00 AM EST:''' The data centre is under annual maintenance. All systems are offline. Systems are expected to be back late afternoon today; check for updates on this page.&lt;br /&gt;
&lt;br /&gt;
'''May 18, 2018:''' Announcement: Annual scheduled maintenance downtime: Thursday May 24, starting 7:00 AM&lt;br /&gt;
&lt;br /&gt;
'''May 16, 2018:''' Cooling  restored, systems online&lt;br /&gt;
&lt;br /&gt;
'''May 16, 2018:''' Cooling issue at datacentre again, all systems down&lt;br /&gt;
&lt;br /&gt;
'''May 15, 2018:''' Cooling restored, systems coming online&lt;br /&gt;
&lt;br /&gt;
'''May 15, 2018''' Cooling issue at datacentre, all systems down&lt;br /&gt;
&lt;br /&gt;
'''May 4, 2018:''' [[HPSS]] is now operational on Niagara.&lt;br /&gt;
&lt;br /&gt;
'''May 3, 2018:''' [[Burst Buffer]] is available upon request.&lt;br /&gt;
&lt;br /&gt;
'''May 3, 2018:''' The [https://docs.computecanada.ca/wiki/Globus Globus] endpoint for Niagara is available: computecanada#niagara.&lt;br /&gt;
&lt;br /&gt;
'''May 1, 2018:''' System status moved he here.&lt;br /&gt;
&lt;br /&gt;
'''Apr 23, 2018:''' GPC-compute is decommissioned, GPC-storage available until 30 May 2018.&lt;br /&gt;
&lt;br /&gt;
'''April 10, 2018:''' Niagara commissioned.&lt;/div&gt;</summary>
		<author><name>Rzon</name></author>
	</entry>
	<entry>
		<id>https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7622</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7622"/>
		<updated>2026-03-25T20:59:41Z</updated>

		<summary type="html">&lt;p&gt;Rzon: /* System Status */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;__NOTOC__&lt;br /&gt;
{| style=&amp;quot;border-spacing:10px; width: 95%&amp;quot;&lt;br /&gt;
| style=&amp;quot;padding:1em; padding-top:.1em; border:2px solid #0645ad; background-color:#f6f6f6; border-radius:7px&amp;quot;|&lt;br /&gt;
&lt;br /&gt;
==System Status==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- Use &amp;quot;Up&amp;quot;, &amp;quot;Partial&amp;quot; or &amp;quot;Down&amp;quot;; these are templates. --&amp;gt;&lt;br /&gt;
{|style=&amp;quot;width:100%&amp;quot; &lt;br /&gt;
|{{Up3 | Trillium|https://docs.alliancecan.ca/wiki/Trillium_Quickstart}}&lt;br /&gt;
|{{Up3 | OnDemand|https://docs.alliancecan.ca/wiki/Trillium_Open_OnDemand_Quickstart}}&lt;br /&gt;
|{{Up | Globus |Globus}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Down | HPSS|HPSS}}&lt;br /&gt;
|{{Up| Balam|Balam}}&lt;br /&gt;
|{{Up | S4H | S4H}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Up | Teach|Teach}}&lt;br /&gt;
|{{Up3 | File system|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Storage}}&lt;br /&gt;
|{{Up3 | External Network|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Logging_in}} &lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Wed Mar 25, 2026, 5:00 pm:''' Trillium is operational again.&lt;br /&gt;
&lt;br /&gt;
'''Wed Mar 25, 2026, 9:00 am:''' Teach is operational again.&lt;br /&gt;
&lt;br /&gt;
'''Tue Mar 24, 2026, 8:45 pm:''' Open OnDemand is operational again.&lt;br /&gt;
&lt;br /&gt;
'''Tue Mar 24, 2026, 1:00 pm:''' External connectivity is back. &lt;br /&gt;
&lt;br /&gt;
'''Tue Mar 24, 2026, 12:05 pm:''' External connectivity to the data centre was lost. &lt;br /&gt;
&lt;br /&gt;
'''Tue Mar 24, 2026, 7:00 am:''' Maintenance has started.&lt;br /&gt;
&lt;br /&gt;
'''Mon Mar 16, 2026. 13:30pm''' Recovering.  Almost all systems are up again. Please resubmit your jobs that crashed. &lt;br /&gt;
&lt;br /&gt;
'''Mon Mar 16, 2026. 12:00pm''' Power glitch at the data centre caused compute nodes to go down.&lt;br /&gt;
&lt;br /&gt;
'''Thu Mar 12, 2026, 4:15 pm''' Connection to Trillium are operational again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Mar 12, 2026, 1:00 pm''' We've had some login issues particularly for Trillium-GPU. We're investigating.&lt;br /&gt;
&lt;br /&gt;
'''Downtime Announcement:'''  The winter cooling tower maintenance for the SciNet data centre will take place on March 24 and 25, 2026, starting at 7:00 a.m. on the 24th.  All SciNet systems (Trillium, OnDemand, Balam, S4H, Teach, as well as hosted equipment) will have their compute nodes shut down. Login nodes, file systems, and the HPSS system will remain available, and&lt;br /&gt;
jobs will be held in the queue until maintenance is complete.  Starting 7am on Mar 23, users are encouraged to submit small and short jobs that may be scheduled before the maintenance begins.&lt;br /&gt;
&lt;br /&gt;
'''Fri Feb 20, 2026, 11:35 pm:''' Power glitch, ~480 compute nodes rebooted. Regional power quality has been quite poor lately ([https://www.yorkregion.com/news/road-salt-blamed-for-power-outages/article_1a36d25d-5f97-56ee-a0c7-c49c7b732d38.html 1],&lt;br /&gt;
[https://www.yorkregion.com/news/power-company-executive-responds-to-york-region-outages/article_c4d072e7-2892-5c9c-8deb-ac5e1936779c.html 2]).&lt;br /&gt;
&lt;br /&gt;
'''Thu Feb 19, 2026, 3:00 pm:''' Systems restored. Please report issues to support@scinet.utoronto.ca.&lt;br /&gt;
&lt;br /&gt;
'''Tue Feb 17, 2026, 8:40 am:''' Power outage at the data centre.  Cooling issues have developed as a result.  Major systems (Trillium, S4H) are expected to be down until sometime Thursday. Login nodes and file systems will remain accessible.&lt;br /&gt;
&lt;br /&gt;
'''Mon Feb 16, 2026, 8:40 pm:''' Electricity is unstable in the data centre area due to severe snowfall.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 1:40 pm:''' All services are operational again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 12:00 pm:''' The Trillium and Open OnDemand compute nodes are operational again. We are still working on bringing Balam, Neptune and S4H nodes up again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 10:00 am:''' There was a power glitch at the data centre overnight. The login nodes are accessible but the compute nodes are down.  &lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--  When removing system status entries, please archive them to: --&amp;gt;&lt;br /&gt;
[[Previous messages]]&lt;br /&gt;
&lt;br /&gt;
{|style=&amp;quot;border-spacing: 10px;width: 100%&amp;quot;&lt;br /&gt;
|valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== QuickStart Guides ==&lt;br /&gt;
* [https://docs.alliancecan.ca/wiki/Trillium_Quickstart Trillium Quickstart]&lt;br /&gt;
* [[HPSS | HPSS archival storage]]&lt;br /&gt;
* [[Teach|Teach cluster]]&lt;br /&gt;
* [[FAQ | FAQ (frequently asked questions)]]&lt;br /&gt;
* [[Acknowledging SciNet]]&lt;br /&gt;
| valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== Tutorials, Manuals, etc. ==&lt;br /&gt;
* [https://education.scinet.utoronto.ca SciNet education material]&lt;br /&gt;
* [https://www.youtube.com/c/SciNetHPCattheUniversityofToronto SciNet's YouTube channel]&lt;br /&gt;
* [[Modules for Mist]] &lt;br /&gt;
* [[Commercial software]]&lt;br /&gt;
* [[Burst Buffer]]&lt;br /&gt;
* [[SSH#SSH Keys|SSH keys]]&lt;br /&gt;
* [[SSH Tunneling]]&lt;br /&gt;
* [[Visualization]]&lt;br /&gt;
* [[Running Serial Jobs on Niagara]]&lt;br /&gt;
* [[Jupyter Hub]]&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Rzon</name></author>
	</entry>
	<entry>
		<id>https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7619</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7619"/>
		<updated>2026-03-25T20:59:28Z</updated>

		<summary type="html">&lt;p&gt;Rzon: /* System Status */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;__NOTOC__&lt;br /&gt;
{| style=&amp;quot;border-spacing:10px; width: 95%&amp;quot;&lt;br /&gt;
| style=&amp;quot;padding:1em; padding-top:.1em; border:2px solid #0645ad; background-color:#f6f6f6; border-radius:7px&amp;quot;|&lt;br /&gt;
&lt;br /&gt;
==System Status==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- Use &amp;quot;Up&amp;quot;, &amp;quot;Partial&amp;quot; or &amp;quot;Down&amp;quot;; these are templates. --&amp;gt;&lt;br /&gt;
{|style=&amp;quot;width:100%&amp;quot; &lt;br /&gt;
|{{Up3 | Trillium|https://docs.alliancecan.ca/wiki/Trillium_Quickstart}}&lt;br /&gt;
|{{Up3 | OnDemand|https://docs.alliancecan.ca/wiki/Trillium_Open_OnDemand_Quickstart}}&lt;br /&gt;
|{{Up | Globus |Globus}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Down | HPSS|HPSS}}&lt;br /&gt;
|{{Partial | Balam|Balam}}&lt;br /&gt;
|{{Up | S4H | S4H}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Up | Teach|Teach}}&lt;br /&gt;
|{{Up3 | File system|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Storage}}&lt;br /&gt;
|{{Up3 | External Network|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Logging_in}} &lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Wed Mar 25, 2026, 5:00 pm:''' Trillium is operational again.&lt;br /&gt;
&lt;br /&gt;
'''Wed Mar 25, 2026, 9:00 am:''' Teach is operational again.&lt;br /&gt;
&lt;br /&gt;
'''Tue Mar 24, 2026, 8:45 pm:''' Open OnDemand is operational again.&lt;br /&gt;
&lt;br /&gt;
'''Tue Mar 24, 2026, 1:00 pm:''' External connectivity is back. &lt;br /&gt;
&lt;br /&gt;
'''Tue Mar 24, 2026, 12:05 pm:''' External connectivity to the data centre was lost. &lt;br /&gt;
&lt;br /&gt;
'''Tue Mar 24, 2026, 7:00 am:''' Maintenance has started.&lt;br /&gt;
&lt;br /&gt;
'''Mon Mar 16, 2026. 13:30pm''' Recovering.  Almost all systems are up again. Please resubmit your jobs that crashed. &lt;br /&gt;
&lt;br /&gt;
'''Mon Mar 16, 2026. 12:00pm''' Power glitch at the data centre caused compute nodes to go down.&lt;br /&gt;
&lt;br /&gt;
'''Thu Mar 12, 2026, 4:15 pm''' Connection to Trillium are operational again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Mar 12, 2026, 1:00 pm''' We've had some login issues particularly for Trillium-GPU. We're investigating.&lt;br /&gt;
&lt;br /&gt;
'''Downtime Announcement:'''  The winter cooling tower maintenance for the SciNet data centre will take place on March 24 and 25, 2026, starting at 7:00 a.m. on the 24th.  All SciNet systems (Trillium, OnDemand, Balam, S4H, Teach, as well as hosted equipment) will have their compute nodes shut down. Login nodes, file systems, and the HPSS system will remain available, and&lt;br /&gt;
jobs will be held in the queue until maintenance is complete.  Starting 7am on Mar 23, users are encouraged to submit small and short jobs that may be scheduled before the maintenance begins.&lt;br /&gt;
&lt;br /&gt;
'''Fri Feb 20, 2026, 11:35 pm:''' Power glitch, ~480 compute nodes rebooted. Regional power quality has been quite poor lately ([https://www.yorkregion.com/news/road-salt-blamed-for-power-outages/article_1a36d25d-5f97-56ee-a0c7-c49c7b732d38.html 1],&lt;br /&gt;
[https://www.yorkregion.com/news/power-company-executive-responds-to-york-region-outages/article_c4d072e7-2892-5c9c-8deb-ac5e1936779c.html 2]).&lt;br /&gt;
&lt;br /&gt;
'''Thu Feb 19, 2026, 3:00 pm:''' Systems restored. Please report issues to support@scinet.utoronto.ca.&lt;br /&gt;
&lt;br /&gt;
'''Tue Feb 17, 2026, 8:40 am:''' Power outage at the data centre.  Cooling issues have developed as a result.  Major systems (Trillium, S4H) are expected to be down until sometime Thursday. Login nodes and file systems will remain accessible.&lt;br /&gt;
&lt;br /&gt;
'''Mon Feb 16, 2026, 8:40 pm:''' Electricity is unstable in the data centre area due to severe snowfall.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 1:40 pm:''' All services are operational again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 12:00 pm:''' The Trillium and Open OnDemand compute nodes are operational again. We are still working on bringing Balam, Neptune and S4H nodes up again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 10:00 am:''' There was a power glitch at the data centre overnight. The login nodes are accessible but the compute nodes are down.  &lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--  When removing system status entries, please archive them to: --&amp;gt;&lt;br /&gt;
[[Previous messages]]&lt;br /&gt;
&lt;br /&gt;
{|style=&amp;quot;border-spacing: 10px;width: 100%&amp;quot;&lt;br /&gt;
|valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== QuickStart Guides ==&lt;br /&gt;
* [https://docs.alliancecan.ca/wiki/Trillium_Quickstart Trillium Quickstart]&lt;br /&gt;
* [[HPSS | HPSS archival storage]]&lt;br /&gt;
* [[Teach|Teach cluster]]&lt;br /&gt;
* [[FAQ | FAQ (frequently asked questions)]]&lt;br /&gt;
* [[Acknowledging SciNet]]&lt;br /&gt;
| valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== Tutorials, Manuals, etc. ==&lt;br /&gt;
* [https://education.scinet.utoronto.ca SciNet education material]&lt;br /&gt;
* [https://www.youtube.com/c/SciNetHPCattheUniversityofToronto SciNet's YouTube channel]&lt;br /&gt;
* [[Modules for Mist]] &lt;br /&gt;
* [[Commercial software]]&lt;br /&gt;
* [[Burst Buffer]]&lt;br /&gt;
* [[SSH#SSH Keys|SSH keys]]&lt;br /&gt;
* [[SSH Tunneling]]&lt;br /&gt;
* [[Visualization]]&lt;br /&gt;
* [[Running Serial Jobs on Niagara]]&lt;br /&gt;
* [[Jupyter Hub]]&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Rzon</name></author>
	</entry>
	<entry>
		<id>https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7613</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7613"/>
		<updated>2026-03-25T13:38:46Z</updated>

		<summary type="html">&lt;p&gt;Rzon: /* System Status */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;__NOTOC__&lt;br /&gt;
{| style=&amp;quot;border-spacing:10px; width: 95%&amp;quot;&lt;br /&gt;
| style=&amp;quot;padding:1em; padding-top:.1em; border:2px solid #0645ad; background-color:#f6f6f6; border-radius:7px&amp;quot;|&lt;br /&gt;
&lt;br /&gt;
==System Status==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- Use &amp;quot;Up&amp;quot;, &amp;quot;Partial&amp;quot; or &amp;quot;Down&amp;quot;; these are templates. --&amp;gt;&lt;br /&gt;
{|style=&amp;quot;width:100%&amp;quot; &lt;br /&gt;
|{{Partial3 | Trillium|https://docs.alliancecan.ca/wiki/Trillium_Quickstart}}&lt;br /&gt;
|{{Up3 | OnDemand|https://docs.alliancecan.ca/wiki/Trillium_Open_OnDemand_Quickstart}}&lt;br /&gt;
|{{Up | Globus |Globus}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Down | HPSS|HPSS}}&lt;br /&gt;
|{{Partial | Balam|Balam}}&lt;br /&gt;
|{{Down | S4H | S4H}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Up | Teach|Teach}}&lt;br /&gt;
|{{Up3 | File system|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Storage}}&lt;br /&gt;
|{{Up3 | External Network|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Logging_in}} &lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Wed Mar 25, 2026, 9:00 pm:''' Teach is operational again.&lt;br /&gt;
&lt;br /&gt;
'''Tue Mar 24, 2026, 8:45 pm:''' Open OnDemand is operational again.&lt;br /&gt;
&lt;br /&gt;
'''Tue Mar 24, 2026, 1:00 pm:''' External connectivity is back. &lt;br /&gt;
&lt;br /&gt;
'''Tue Mar 24, 2026, 12:05 pm:''' External connectivity to the data centre was lost. &lt;br /&gt;
&lt;br /&gt;
'''Tue Mar 24, 2026, 7:00 am:''' Maintenance has started.&lt;br /&gt;
&lt;br /&gt;
'''Mon Mar 16, 2026. 13:30pm''' Recovering.  Almost all systems are up again. Please resubmit your jobs that crashed. &lt;br /&gt;
&lt;br /&gt;
'''Mon Mar 16, 2026. 12:00pm''' Power glitch at the data centre caused compute nodes to go down.&lt;br /&gt;
&lt;br /&gt;
'''Thu Mar 12, 2026, 4:15 pm''' Connection to Trillium are operational again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Mar 12, 2026, 1:00 pm''' We've had some login issues particularly for Trillium-GPU. We're investigating.&lt;br /&gt;
&lt;br /&gt;
'''Downtime Announcement:'''  The winter cooling tower maintenance for the SciNet data centre will take place on March 24 and 25, 2026, starting at 7:00 a.m. on the 24th.  All SciNet systems (Trillium, OnDemand, Balam, S4H, Teach, as well as hosted equipment) will have their compute nodes shut down. Login nodes, file systems, and the HPSS system will remain available, and&lt;br /&gt;
jobs will be held in the queue until maintenance is complete.  Starting 7am on Mar 23, users are encouraged to submit small and short jobs that may be scheduled before the maintenance begins.&lt;br /&gt;
&lt;br /&gt;
'''Fri Feb 20, 2026, 11:35 pm:''' Power glitch, ~480 compute nodes rebooted. Regional power quality has been quite poor lately ([https://www.yorkregion.com/news/road-salt-blamed-for-power-outages/article_1a36d25d-5f97-56ee-a0c7-c49c7b732d38.html 1],&lt;br /&gt;
[https://www.yorkregion.com/news/power-company-executive-responds-to-york-region-outages/article_c4d072e7-2892-5c9c-8deb-ac5e1936779c.html 2]).&lt;br /&gt;
&lt;br /&gt;
'''Thu Feb 19, 2026, 3:00 pm:''' Systems restored. Please report issues to support@scinet.utoronto.ca.&lt;br /&gt;
&lt;br /&gt;
'''Tue Feb 17, 2026, 8:40 am:''' Power outage at the data centre.  Cooling issues have developed as a result.  Major systems (Trillium, S4H) are expected to be down until sometime Thursday. Login nodes and file systems will remain accessible.&lt;br /&gt;
&lt;br /&gt;
'''Mon Feb 16, 2026, 8:40 pm:''' Electricity is unstable in the data centre area due to severe snowfall.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 1:40 pm:''' All services are operational again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 12:00 pm:''' The Trillium and Open OnDemand compute nodes are operational again. We are still working on bringing Balam, Neptune and S4H nodes up again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 10:00 am:''' There was a power glitch at the data centre overnight. The login nodes are accessible but the compute nodes are down.  &lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--  When removing system status entries, please archive them to: --&amp;gt;&lt;br /&gt;
[[Previous messages]]&lt;br /&gt;
&lt;br /&gt;
{|style=&amp;quot;border-spacing: 10px;width: 100%&amp;quot;&lt;br /&gt;
|valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== QuickStart Guides ==&lt;br /&gt;
* [https://docs.alliancecan.ca/wiki/Trillium_Quickstart Trillium Quickstart]&lt;br /&gt;
* [[HPSS | HPSS archival storage]]&lt;br /&gt;
* [[Teach|Teach cluster]]&lt;br /&gt;
* [[FAQ | FAQ (frequently asked questions)]]&lt;br /&gt;
* [[Acknowledging SciNet]]&lt;br /&gt;
| valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== Tutorials, Manuals, etc. ==&lt;br /&gt;
* [https://education.scinet.utoronto.ca SciNet education material]&lt;br /&gt;
* [https://www.youtube.com/c/SciNetHPCattheUniversityofToronto SciNet's YouTube channel]&lt;br /&gt;
* [[Modules for Mist]] &lt;br /&gt;
* [[Commercial software]]&lt;br /&gt;
* [[Burst Buffer]]&lt;br /&gt;
* [[SSH#SSH Keys|SSH keys]]&lt;br /&gt;
* [[SSH Tunneling]]&lt;br /&gt;
* [[Visualization]]&lt;br /&gt;
* [[Running Serial Jobs on Niagara]]&lt;br /&gt;
* [[Jupyter Hub]]&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Rzon</name></author>
	</entry>
	<entry>
		<id>https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7610</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7610"/>
		<updated>2026-03-25T13:31:11Z</updated>

		<summary type="html">&lt;p&gt;Rzon: /* System Status */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;__NOTOC__&lt;br /&gt;
{| style=&amp;quot;border-spacing:10px; width: 95%&amp;quot;&lt;br /&gt;
| style=&amp;quot;padding:1em; padding-top:.1em; border:2px solid #0645ad; background-color:#f6f6f6; border-radius:7px&amp;quot;|&lt;br /&gt;
&lt;br /&gt;
==System Status==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- Use &amp;quot;Up&amp;quot;, &amp;quot;Partial&amp;quot; or &amp;quot;Down&amp;quot;; these are templates. --&amp;gt;&lt;br /&gt;
{|style=&amp;quot;width:100%&amp;quot; &lt;br /&gt;
|{{Partial3 | Trillium|https://docs.alliancecan.ca/wiki/Trillium_Quickstart}}&lt;br /&gt;
|{{Up3 | OnDemand|https://docs.alliancecan.ca/wiki/Trillium_Open_OnDemand_Quickstart}}&lt;br /&gt;
|{{Up | Globus |Globus}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Down | HPSS|HPSS}}&lt;br /&gt;
|{{Down | Balam|Balam}}&lt;br /&gt;
|{{Down | S4H | S4H}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Up | Teach|Teach}}&lt;br /&gt;
|{{Up3 | File system|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Storage}}&lt;br /&gt;
|{{Up3 | External Network|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Logging_in}} &lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Wed Mar 25, 2026, 9:00 pm:''' Teach is operational again.&lt;br /&gt;
&lt;br /&gt;
'''Tue Mar 24, 2026, 8:45 pm:''' Open OnDemand is operational again.&lt;br /&gt;
&lt;br /&gt;
'''Tue Mar 24, 2026, 1:00 pm:''' External connectivity is back. &lt;br /&gt;
&lt;br /&gt;
'''Tue Mar 24, 2026, 12:05 pm:''' External connectivity to the data centre was lost. &lt;br /&gt;
&lt;br /&gt;
'''Tue Mar 24, 2026, 7:00 am:''' Maintenance has started.&lt;br /&gt;
&lt;br /&gt;
'''Mon Mar 16, 2026. 13:30pm''' Recovering.  Almost all systems are up again. Please resubmit your jobs that crashed. &lt;br /&gt;
&lt;br /&gt;
'''Mon Mar 16, 2026. 12:00pm''' Power glitch at the data centre caused compute nodes to go down.&lt;br /&gt;
&lt;br /&gt;
'''Thu Mar 12, 2026, 4:15 pm''' Connection to Trillium are operational again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Mar 12, 2026, 1:00 pm''' We've had some login issues particularly for Trillium-GPU. We're investigating.&lt;br /&gt;
&lt;br /&gt;
'''Downtime Announcement:'''  The winter cooling tower maintenance for the SciNet data centre will take place on March 24 and 25, 2026, starting at 7:00 a.m. on the 24th.  All SciNet systems (Trillium, OnDemand, Balam, S4H, Teach, as well as hosted equipment) will have their compute nodes shut down. Login nodes, file systems, and the HPSS system will remain available, and&lt;br /&gt;
jobs will be held in the queue until maintenance is complete.  Starting 7am on Mar 23, users are encouraged to submit small and short jobs that may be scheduled before the maintenance begins.&lt;br /&gt;
&lt;br /&gt;
'''Fri Feb 20, 2026, 11:35 pm:''' Power glitch, ~480 compute nodes rebooted. Regional power quality has been quite poor lately ([https://www.yorkregion.com/news/road-salt-blamed-for-power-outages/article_1a36d25d-5f97-56ee-a0c7-c49c7b732d38.html 1],&lt;br /&gt;
[https://www.yorkregion.com/news/power-company-executive-responds-to-york-region-outages/article_c4d072e7-2892-5c9c-8deb-ac5e1936779c.html 2]).&lt;br /&gt;
&lt;br /&gt;
'''Thu Feb 19, 2026, 3:00 pm:''' Systems restored. Please report issues to support@scinet.utoronto.ca.&lt;br /&gt;
&lt;br /&gt;
'''Tue Feb 17, 2026, 8:40 am:''' Power outage at the data centre.  Cooling issues have developed as a result.  Major systems (Trillium, S4H) are expected to be down until sometime Thursday. Login nodes and file systems will remain accessible.&lt;br /&gt;
&lt;br /&gt;
'''Mon Feb 16, 2026, 8:40 pm:''' Electricity is unstable in the data centre area due to severe snowfall.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 1:40 pm:''' All services are operational again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 12:00 pm:''' The Trillium and Open OnDemand compute nodes are operational again. We are still working on bringing Balam, Neptune and S4H nodes up again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 10:00 am:''' There was a power glitch at the data centre overnight. The login nodes are accessible but the compute nodes are down.  &lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--  When removing system status entries, please archive them to: --&amp;gt;&lt;br /&gt;
[[Previous messages]]&lt;br /&gt;
&lt;br /&gt;
{|style=&amp;quot;border-spacing: 10px;width: 100%&amp;quot;&lt;br /&gt;
|valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== QuickStart Guides ==&lt;br /&gt;
* [https://docs.alliancecan.ca/wiki/Trillium_Quickstart Trillium Quickstart]&lt;br /&gt;
* [[HPSS | HPSS archival storage]]&lt;br /&gt;
* [[Teach|Teach cluster]]&lt;br /&gt;
* [[FAQ | FAQ (frequently asked questions)]]&lt;br /&gt;
* [[Acknowledging SciNet]]&lt;br /&gt;
| valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== Tutorials, Manuals, etc. ==&lt;br /&gt;
* [https://education.scinet.utoronto.ca SciNet education material]&lt;br /&gt;
* [https://www.youtube.com/c/SciNetHPCattheUniversityofToronto SciNet's YouTube channel]&lt;br /&gt;
* [[Modules for Mist]] &lt;br /&gt;
* [[Commercial software]]&lt;br /&gt;
* [[Burst Buffer]]&lt;br /&gt;
* [[SSH#SSH Keys|SSH keys]]&lt;br /&gt;
* [[SSH Tunneling]]&lt;br /&gt;
* [[Visualization]]&lt;br /&gt;
* [[Running Serial Jobs on Niagara]]&lt;br /&gt;
* [[Jupyter Hub]]&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Rzon</name></author>
	</entry>
	<entry>
		<id>https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7604</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7604"/>
		<updated>2026-03-24T17:20:04Z</updated>

		<summary type="html">&lt;p&gt;Rzon: /* System Status */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;__NOTOC__&lt;br /&gt;
{| style=&amp;quot;border-spacing:10px; width: 95%&amp;quot;&lt;br /&gt;
| style=&amp;quot;padding:1em; padding-top:.1em; border:2px solid #0645ad; background-color:#f6f6f6; border-radius:7px&amp;quot;|&lt;br /&gt;
&lt;br /&gt;
==System Status==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- Use &amp;quot;Up&amp;quot;, &amp;quot;Partial&amp;quot; or &amp;quot;Down&amp;quot;; these are templates. --&amp;gt;&lt;br /&gt;
{|style=&amp;quot;width:100%&amp;quot; &lt;br /&gt;
|{{Partial3 | Trillium|https://docs.alliancecan.ca/wiki/Trillium_Quickstart}}&lt;br /&gt;
|{{Partial3 | OnDemand|https://docs.alliancecan.ca/wiki/Trillium_Open_OnDemand_Quickstart}}&lt;br /&gt;
|{{Up | Globus |Globus}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Down | HPSS|HPSS}}&lt;br /&gt;
|{{Down | Balam|Balam}}&lt;br /&gt;
|{{Down | S4H | S4H}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Partial | Teach|Teach}}&lt;br /&gt;
|{{Up3 | File system|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Storage}}&lt;br /&gt;
|{{Up3 | External Network|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Logging_in}} &lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Tue Mar 24, 2026, 1:00 pm:''' External connectivity is back. &lt;br /&gt;
&lt;br /&gt;
'''Tue Mar 24, 2026, 12:05 pm:''' External connectivity to the data centre was lost. &lt;br /&gt;
&lt;br /&gt;
'''Tue Mar 24, 2026, 7:00 am:''' Maintenance has started.&lt;br /&gt;
&lt;br /&gt;
'''Mon Mar 16, 2026. 13:30pm''' Recovering.  Almost all systems are up again. Please resubmit your jobs that crashed. &lt;br /&gt;
&lt;br /&gt;
'''Mon Mar 16, 2026. 12:00pm''' Power glitch at the data centre caused compute nodes to go down.&lt;br /&gt;
&lt;br /&gt;
'''Thu Mar 12, 2026, 4:15 pm''' Connection to Trillium are operational again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Mar 12, 2026, 1:00 pm''' We've had some login issues particularly for Trillium-GPU. We're investigating.&lt;br /&gt;
&lt;br /&gt;
'''Downtime Announcement:'''  The winter cooling tower maintenance for the SciNet data centre will take place on March 24 and 25, 2026, starting at 7:00 a.m. on the 24th.  All SciNet systems (Trillium, OnDemand, Balam, S4H, Teach, as well as hosted equipment) will have their compute nodes shut down. Login nodes, file systems, and the HPSS system will remain available, and&lt;br /&gt;
jobs will be held in the queue until maintenance is complete.  Starting 7am on Mar 23, users are encouraged to submit small and short jobs that may be scheduled before the maintenance begins.&lt;br /&gt;
&lt;br /&gt;
'''Fri Feb 20, 2026, 11:35 pm:''' Power glitch, ~480 compute nodes rebooted. Regional power quality has been quite poor lately ([https://www.yorkregion.com/news/road-salt-blamed-for-power-outages/article_1a36d25d-5f97-56ee-a0c7-c49c7b732d38.html 1],&lt;br /&gt;
[https://www.yorkregion.com/news/power-company-executive-responds-to-york-region-outages/article_c4d072e7-2892-5c9c-8deb-ac5e1936779c.html 2]).&lt;br /&gt;
&lt;br /&gt;
'''Thu Feb 19, 2026, 3:00 pm:''' Systems restored. Please report issues to support@scinet.utoronto.ca.&lt;br /&gt;
&lt;br /&gt;
'''Tue Feb 17, 2026, 8:40 am:''' Power outage at the data centre.  Cooling issues have developed as a result.  Major systems (Trillium, S4H) are expected to be down until sometime Thursday. Login nodes and file systems will remain accessible.&lt;br /&gt;
&lt;br /&gt;
'''Mon Feb 16, 2026, 8:40 pm:''' Electricity is unstable in the data centre area due to severe snowfall.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 1:40 pm:''' All services are operational again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 12:00 pm:''' The Trillium and Open OnDemand compute nodes are operational again. We are still working on bringing Balam, Neptune and S4H nodes up again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 10:00 am:''' There was a power glitch at the data centre overnight. The login nodes are accessible but the compute nodes are down.  &lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--  When removing system status entries, please archive them to: --&amp;gt;&lt;br /&gt;
[[Previous messages]]&lt;br /&gt;
&lt;br /&gt;
{|style=&amp;quot;border-spacing: 10px;width: 100%&amp;quot;&lt;br /&gt;
|valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== QuickStart Guides ==&lt;br /&gt;
* [https://docs.alliancecan.ca/wiki/Trillium_Quickstart Trillium Quickstart]&lt;br /&gt;
* [[HPSS | HPSS archival storage]]&lt;br /&gt;
* [[Teach|Teach cluster]]&lt;br /&gt;
* [[FAQ | FAQ (frequently asked questions)]]&lt;br /&gt;
* [[Acknowledging SciNet]]&lt;br /&gt;
| valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== Tutorials, Manuals, etc. ==&lt;br /&gt;
* [https://education.scinet.utoronto.ca SciNet education material]&lt;br /&gt;
* [https://www.youtube.com/c/SciNetHPCattheUniversityofToronto SciNet's YouTube channel]&lt;br /&gt;
* [[Modules for Mist]] &lt;br /&gt;
* [[Commercial software]]&lt;br /&gt;
* [[Burst Buffer]]&lt;br /&gt;
* [[SSH#SSH Keys|SSH keys]]&lt;br /&gt;
* [[SSH Tunneling]]&lt;br /&gt;
* [[Visualization]]&lt;br /&gt;
* [[Running Serial Jobs on Niagara]]&lt;br /&gt;
* [[Jupyter Hub]]&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Rzon</name></author>
	</entry>
	<entry>
		<id>https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7601</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7601"/>
		<updated>2026-03-24T16:24:58Z</updated>

		<summary type="html">&lt;p&gt;Rzon: /* System Status */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;__NOTOC__&lt;br /&gt;
{| style=&amp;quot;border-spacing:10px; width: 95%&amp;quot;&lt;br /&gt;
| style=&amp;quot;padding:1em; padding-top:.1em; border:2px solid #0645ad; background-color:#f6f6f6; border-radius:7px&amp;quot;|&lt;br /&gt;
&lt;br /&gt;
==System Status==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- Use &amp;quot;Up&amp;quot;, &amp;quot;Partial&amp;quot; or &amp;quot;Down&amp;quot;; these are templates. --&amp;gt;&lt;br /&gt;
{|style=&amp;quot;width:100%&amp;quot; &lt;br /&gt;
|{{Partial3 | Trillium|https://docs.alliancecan.ca/wiki/Trillium_Quickstart}}&lt;br /&gt;
|{{Partial3 | OnDemand|https://docs.alliancecan.ca/wiki/Trillium_Open_OnDemand_Quickstart}}&lt;br /&gt;
|{{Up | Globus |Globus}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Down | HPSS|HPSS}}&lt;br /&gt;
|{{Down | Balam|Balam}}&lt;br /&gt;
|{{Down | S4H | S4H}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Partial | Teach|Teach}}&lt;br /&gt;
|{{Up3 | File system|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Storage}}&lt;br /&gt;
|{{Down3 | External Network|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Logging_in}} &lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Tue Mar 24, 2026, 12:05 pm:''' External connectivity to the data centre was lost. &lt;br /&gt;
&lt;br /&gt;
'''Tue Mar 24, 2026, 7:00 am:''' Maintenance has started.&lt;br /&gt;
&lt;br /&gt;
'''Mon Mar 16, 2026. 13:30pm''' Recovering.  Almost all systems are up again. Please resubmit your jobs that crashed. &lt;br /&gt;
&lt;br /&gt;
'''Mon Mar 16, 2026. 12:00pm''' Power glitch at the data centre caused compute nodes to go down.&lt;br /&gt;
&lt;br /&gt;
'''Thu Mar 12, 2026, 4:15 pm''' Connection to Trillium are operational again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Mar 12, 2026, 1:00 pm''' We've had some login issues particularly for Trillium-GPU. We're investigating.&lt;br /&gt;
&lt;br /&gt;
'''Downtime Announcement:'''  The winter cooling tower maintenance for the SciNet data centre will take place on March 24 and 25, 2026, starting at 7:00 a.m. on the 24th.  All SciNet systems (Trillium, OnDemand, Balam, S4H, Teach, as well as hosted equipment) will have their compute nodes shut down. Login nodes, file systems, and the HPSS system will remain available, and&lt;br /&gt;
jobs will be held in the queue until maintenance is complete.  Starting 7am on Mar 23, users are encouraged to submit small and short jobs that may be scheduled before the maintenance begins.&lt;br /&gt;
&lt;br /&gt;
'''Fri Feb 20, 2026, 11:35 pm:''' Power glitch, ~480 compute nodes rebooted. Regional power quality has been quite poor lately ([https://www.yorkregion.com/news/road-salt-blamed-for-power-outages/article_1a36d25d-5f97-56ee-a0c7-c49c7b732d38.html 1],&lt;br /&gt;
[https://www.yorkregion.com/news/power-company-executive-responds-to-york-region-outages/article_c4d072e7-2892-5c9c-8deb-ac5e1936779c.html 2]).&lt;br /&gt;
&lt;br /&gt;
'''Thu Feb 19, 2026, 3:00 pm:''' Systems restored. Please report issues to support@scinet.utoronto.ca.&lt;br /&gt;
&lt;br /&gt;
'''Tue Feb 17, 2026, 8:40 am:''' Power outage at the data centre.  Cooling issues have developed as a result.  Major systems (Trillium, S4H) are expected to be down until sometime Thursday. Login nodes and file systems will remain accessible.&lt;br /&gt;
&lt;br /&gt;
'''Mon Feb 16, 2026, 8:40 pm:''' Electricity is unstable in the data centre area due to severe snowfall.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 1:40 pm:''' All services are operational again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 12:00 pm:''' The Trillium and Open OnDemand compute nodes are operational again. We are still working on bringing Balam, Neptune and S4H nodes up again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 10:00 am:''' There was a power glitch at the data centre overnight. The login nodes are accessible but the compute nodes are down.  &lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--  When removing system status entries, please archive them to: --&amp;gt;&lt;br /&gt;
[[Previous messages]]&lt;br /&gt;
&lt;br /&gt;
{|style=&amp;quot;border-spacing: 10px;width: 100%&amp;quot;&lt;br /&gt;
|valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== QuickStart Guides ==&lt;br /&gt;
* [https://docs.alliancecan.ca/wiki/Trillium_Quickstart Trillium Quickstart]&lt;br /&gt;
* [[HPSS | HPSS archival storage]]&lt;br /&gt;
* [[Teach|Teach cluster]]&lt;br /&gt;
* [[FAQ | FAQ (frequently asked questions)]]&lt;br /&gt;
* [[Acknowledging SciNet]]&lt;br /&gt;
| valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== Tutorials, Manuals, etc. ==&lt;br /&gt;
* [https://education.scinet.utoronto.ca SciNet education material]&lt;br /&gt;
* [https://www.youtube.com/c/SciNetHPCattheUniversityofToronto SciNet's YouTube channel]&lt;br /&gt;
* [[Modules for Mist]] &lt;br /&gt;
* [[Commercial software]]&lt;br /&gt;
* [[Burst Buffer]]&lt;br /&gt;
* [[SSH#SSH Keys|SSH keys]]&lt;br /&gt;
* [[SSH Tunneling]]&lt;br /&gt;
* [[Visualization]]&lt;br /&gt;
* [[Running Serial Jobs on Niagara]]&lt;br /&gt;
* [[Jupyter Hub]]&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Rzon</name></author>
	</entry>
	<entry>
		<id>https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7595</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7595"/>
		<updated>2026-03-24T11:41:39Z</updated>

		<summary type="html">&lt;p&gt;Rzon: /* System Status */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;__NOTOC__&lt;br /&gt;
{| style=&amp;quot;border-spacing:10px; width: 95%&amp;quot;&lt;br /&gt;
| style=&amp;quot;padding:1em; padding-top:.1em; border:2px solid #0645ad; background-color:#f6f6f6; border-radius:7px&amp;quot;|&lt;br /&gt;
&lt;br /&gt;
==System Status==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- Use &amp;quot;Up&amp;quot;, &amp;quot;Partial&amp;quot; or &amp;quot;Down&amp;quot;; these are templates. --&amp;gt;&lt;br /&gt;
{|style=&amp;quot;width:100%&amp;quot; &lt;br /&gt;
|{{Partial3 | Trillium|https://docs.alliancecan.ca/wiki/Trillium_Quickstart}}&lt;br /&gt;
|{{Partial3 | OnDemand|https://docs.alliancecan.ca/wiki/Trillium_Open_OnDemand_Quickstart}}&lt;br /&gt;
|{{Up | Globus |Globus}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Down | HPSS|HPSS}}&lt;br /&gt;
|{{Down | Balam|Balam}}&lt;br /&gt;
|{{Partial | S4H | S4H}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Partial | Teach|Teach}}&lt;br /&gt;
|{{Up3 | File system|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Storage}}&lt;br /&gt;
|{{Up3 | External Network|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Logging_in}} &lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Tue Mar 24, 2026, 7:00 am:''' Maintenance has started.&lt;br /&gt;
&lt;br /&gt;
'''Mon Mar 16, 2026. 13:30pm''' Recovering.  Almost all systems are up again. Please resubmit your jobs that crashed. &lt;br /&gt;
&lt;br /&gt;
'''Mon Mar 16, 2026. 12:00pm''' Power glitch at the data centre caused compute nodes to go down.&lt;br /&gt;
&lt;br /&gt;
'''Thu Mar 12, 2026, 4:15 pm''' Connection to Trillium are operational again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Mar 12, 2026, 1:00 pm''' We've had some login issues particularly for Trillium-GPU. We're investigating.&lt;br /&gt;
&lt;br /&gt;
'''Downtime Announcement:'''  The winter cooling tower maintenance for the SciNet data centre will take place on March 24 and 25, 2026, starting at 7:00 a.m. on the 24th.  All SciNet systems (Trillium, OnDemand, Balam, S4H, Teach, as well as hosted equipment) will have their compute nodes shut down. Login nodes, file systems, and the HPSS system will remain available, and&lt;br /&gt;
jobs will be held in the queue until maintenance is complete.  Starting 7am on Mar 23, users are encouraged to submit small and short jobs that may be scheduled before the maintenance begins.&lt;br /&gt;
&lt;br /&gt;
'''Fri Feb 20, 2026, 11:35 pm:''' Power glitch, ~480 compute nodes rebooted. Regional power quality has been quite poor lately ([https://www.yorkregion.com/news/road-salt-blamed-for-power-outages/article_1a36d25d-5f97-56ee-a0c7-c49c7b732d38.html 1],&lt;br /&gt;
[https://www.yorkregion.com/news/power-company-executive-responds-to-york-region-outages/article_c4d072e7-2892-5c9c-8deb-ac5e1936779c.html 2]).&lt;br /&gt;
&lt;br /&gt;
'''Thu Feb 19, 2026, 3:00 pm:''' Systems restored. Please report issues to support@scinet.utoronto.ca.&lt;br /&gt;
&lt;br /&gt;
'''Tue Feb 17, 2026, 8:40 am:''' Power outage at the data centre.  Cooling issues have developed as a result.  Major systems (Trillium, S4H) are expected to be down until sometime Thursday. Login nodes and file systems will remain accessible.&lt;br /&gt;
&lt;br /&gt;
'''Mon Feb 16, 2026, 8:40 pm:''' Electricity is unstable in the data centre area due to severe snowfall.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 1:40 pm:''' All services are operational again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 12:00 pm:''' The Trillium and Open OnDemand compute nodes are operational again. We are still working on bringing Balam, Neptune and S4H nodes up again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 10:00 am:''' There was a power glitch at the data centre overnight. The login nodes are accessible but the compute nodes are down.  &lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--  When removing system status entries, please archive them to: --&amp;gt;&lt;br /&gt;
[[Previous messages]]&lt;br /&gt;
&lt;br /&gt;
{|style=&amp;quot;border-spacing: 10px;width: 100%&amp;quot;&lt;br /&gt;
|valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== QuickStart Guides ==&lt;br /&gt;
* [https://docs.alliancecan.ca/wiki/Trillium_Quickstart Trillium Quickstart]&lt;br /&gt;
* [[HPSS | HPSS archival storage]]&lt;br /&gt;
* [[Teach|Teach cluster]]&lt;br /&gt;
* [[FAQ | FAQ (frequently asked questions)]]&lt;br /&gt;
* [[Acknowledging SciNet]]&lt;br /&gt;
| valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== Tutorials, Manuals, etc. ==&lt;br /&gt;
* [https://education.scinet.utoronto.ca SciNet education material]&lt;br /&gt;
* [https://www.youtube.com/c/SciNetHPCattheUniversityofToronto SciNet's YouTube channel]&lt;br /&gt;
* [[Modules for Mist]] &lt;br /&gt;
* [[Commercial software]]&lt;br /&gt;
* [[Burst Buffer]]&lt;br /&gt;
* [[SSH#SSH Keys|SSH keys]]&lt;br /&gt;
* [[SSH Tunneling]]&lt;br /&gt;
* [[Visualization]]&lt;br /&gt;
* [[Running Serial Jobs on Niagara]]&lt;br /&gt;
* [[Jupyter Hub]]&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Rzon</name></author>
	</entry>
	<entry>
		<id>https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7592</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7592"/>
		<updated>2026-03-17T14:24:36Z</updated>

		<summary type="html">&lt;p&gt;Rzon: /* System Status */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;__NOTOC__&lt;br /&gt;
{| style=&amp;quot;border-spacing:10px; width: 95%&amp;quot;&lt;br /&gt;
| style=&amp;quot;padding:1em; padding-top:.1em; border:2px solid #0645ad; background-color:#f6f6f6; border-radius:7px&amp;quot;|&lt;br /&gt;
&lt;br /&gt;
==System Status==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- Use &amp;quot;Up&amp;quot;, &amp;quot;Partial&amp;quot; or &amp;quot;Down&amp;quot;; these are templates. --&amp;gt;&lt;br /&gt;
{|style=&amp;quot;width:100%&amp;quot; &lt;br /&gt;
|{{Up3 | Trillium|https://docs.alliancecan.ca/wiki/Trillium_Quickstart}}&lt;br /&gt;
|{{Up3 | OnDemand|https://docs.alliancecan.ca/wiki/Trillium_Open_OnDemand_Quickstart}}&lt;br /&gt;
|{{Up | Globus |Globus}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Up | HPSS|HPSS}}&lt;br /&gt;
|{{Up | Balam|Balam}}&lt;br /&gt;
|{{Up | S4H | S4H}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Up | Teach|Teach}}&lt;br /&gt;
|{{Up3 | File system|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Storage}}&lt;br /&gt;
|{{Up3 | External Network|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Logging_in}} &lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Mon Mar 16, 2026. 13:30pm''' Recovering.  Almost all systems are up again. Please resubmit your jobs that crashed. &lt;br /&gt;
&lt;br /&gt;
'''Mon Mar 16, 2026. 12:00pm''' Power glitch at the data centre caused compute nodes to go down.&lt;br /&gt;
&lt;br /&gt;
'''Thu Mar 12, 2026, 4:15 pm''' Connection to Trillium are operational again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Mar 12, 2026, 1:00 pm''' We've had some login issues particularly for Trillium-GPU. We're investigating.&lt;br /&gt;
&lt;br /&gt;
'''Downtime Announcement:'''  The winter cooling tower maintenance for the SciNet data centre will take place on March 24 and 25, 2026, starting at 7:00 a.m. on the 24th.  All SciNet systems (Trillium, OnDemand, Balam, S4H, Teach, as well as hosted equipment) will have their compute nodes shut down. Login nodes, file systems, and the HPSS system will remain available, and&lt;br /&gt;
jobs will be held in the queue until maintenance is complete.  Starting 7am on Mar 23, users are encouraged to submit small and short jobs that may be scheduled before the maintenance begins.&lt;br /&gt;
&lt;br /&gt;
'''Fri Feb 20, 2026, 11:35 pm:''' Power glitch, ~480 compute nodes rebooted. Regional power quality has been quite poor lately ([https://www.yorkregion.com/news/road-salt-blamed-for-power-outages/article_1a36d25d-5f97-56ee-a0c7-c49c7b732d38.html 1],&lt;br /&gt;
[https://www.yorkregion.com/news/power-company-executive-responds-to-york-region-outages/article_c4d072e7-2892-5c9c-8deb-ac5e1936779c.html 2]).&lt;br /&gt;
&lt;br /&gt;
'''Thu Feb 19, 2026, 3:00 pm:''' Systems restored. Please report issues to support@scinet.utoronto.ca.&lt;br /&gt;
&lt;br /&gt;
'''Tue Feb 17, 2026, 8:40 am:''' Power outage at the data centre.  Cooling issues have developed as a result.  Major systems (Trillium, S4H) are expected to be down until sometime Thursday. Login nodes and file systems will remain accessible.&lt;br /&gt;
&lt;br /&gt;
'''Mon Feb 16, 2026, 8:40 pm:''' Electricity is unstable in the data centre area due to severe snowfall.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 1:40 pm:''' All services are operational again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 12:00 pm:''' The Trillium and Open OnDemand compute nodes are operational again. We are still working on bringing Balam, Neptune and S4H nodes up again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 10:00 am:''' There was a power glitch at the data centre overnight. The login nodes are accessible but the compute nodes are down.  &lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--  When removing system status entries, please archive them to: --&amp;gt;&lt;br /&gt;
[[Previous messages]]&lt;br /&gt;
&lt;br /&gt;
{|style=&amp;quot;border-spacing: 10px;width: 100%&amp;quot;&lt;br /&gt;
|valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== QuickStart Guides ==&lt;br /&gt;
* [https://docs.alliancecan.ca/wiki/Trillium_Quickstart Trillium Quickstart]&lt;br /&gt;
* [[HPSS | HPSS archival storage]]&lt;br /&gt;
* [[Teach|Teach cluster]]&lt;br /&gt;
* [[FAQ | FAQ (frequently asked questions)]]&lt;br /&gt;
* [[Acknowledging SciNet]]&lt;br /&gt;
| valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== Tutorials, Manuals, etc. ==&lt;br /&gt;
* [https://education.scinet.utoronto.ca SciNet education material]&lt;br /&gt;
* [https://www.youtube.com/c/SciNetHPCattheUniversityofToronto SciNet's YouTube channel]&lt;br /&gt;
* [[Modules for Mist]] &lt;br /&gt;
* [[Commercial software]]&lt;br /&gt;
* [[Burst Buffer]]&lt;br /&gt;
* [[SSH#SSH Keys|SSH keys]]&lt;br /&gt;
* [[SSH Tunneling]]&lt;br /&gt;
* [[Visualization]]&lt;br /&gt;
* [[Running Serial Jobs on Niagara]]&lt;br /&gt;
* [[Jupyter Hub]]&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Rzon</name></author>
	</entry>
	<entry>
		<id>https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7589</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7589"/>
		<updated>2026-03-16T18:07:06Z</updated>

		<summary type="html">&lt;p&gt;Rzon: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;__NOTOC__&lt;br /&gt;
{| style=&amp;quot;border-spacing:10px; width: 95%&amp;quot;&lt;br /&gt;
| style=&amp;quot;padding:1em; padding-top:.1em; border:2px solid #0645ad; background-color:#f6f6f6; border-radius:7px&amp;quot;|&lt;br /&gt;
&lt;br /&gt;
==System Status==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- Use &amp;quot;Up&amp;quot;, &amp;quot;Partial&amp;quot; or &amp;quot;Down&amp;quot;; these are templates. --&amp;gt;&lt;br /&gt;
{|style=&amp;quot;width:100%&amp;quot; &lt;br /&gt;
|{{Partial3 | Trillium|https://docs.alliancecan.ca/wiki/Trillium_Quickstart}}&lt;br /&gt;
|{{Up3 | OnDemand|https://docs.alliancecan.ca/wiki/Trillium_Open_OnDemand_Quickstart}}&lt;br /&gt;
|{{Up | Globus |Globus}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Up | HPSS|HPSS}}&lt;br /&gt;
|{{Up | Balam|Balam}}&lt;br /&gt;
|{{Up | S4H | S4H}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Up | Teach|Teach}}&lt;br /&gt;
|{{Up3 | File system|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Storage}}&lt;br /&gt;
|{{Up3 | External Network|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Logging_in}} &lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Mon Mar 16, 2026. 13:30pm''' Recover.  Almost all systems are up again. Please resubmit your jobs that crashed. &lt;br /&gt;
&lt;br /&gt;
'''Mon Mar 16, 2026. 12:00pm''' Power glitch at the data centre caused compute nodes to go down.&lt;br /&gt;
&lt;br /&gt;
'''Thu Mar 12, 2026, 4:15 pm''' Connection to Trillium are operational again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Mar 12, 2026, 1:00 pm''' We've had some login issues particularly for Trillium-GPU. We're investigating.&lt;br /&gt;
&lt;br /&gt;
'''Downtime Announcement:'''  The winter cooling tower maintenance for the SciNet data centre will take place on March 24 and 25, 2026, starting at 7:00 a.m. on the 24th.  All SciNet systems (Trillium, OnDemand, Balam, S4H, Teach, as well as hosted equipment) will have their compute nodes shut down. Login nodes, file systems, and the HPSS system will remain available, and&lt;br /&gt;
jobs will be held in the queue until maintenance is complete.  Starting 7am on Mar 23, users are encouraged to submit small and short jobs that may be scheduled before the maintenance begins.&lt;br /&gt;
&lt;br /&gt;
'''Fri Feb 20, 2026, 11:35 pm:''' Power glitch, ~480 compute nodes rebooted. Regional power quality has been quite poor lately ([https://www.yorkregion.com/news/road-salt-blamed-for-power-outages/article_1a36d25d-5f97-56ee-a0c7-c49c7b732d38.html 1],&lt;br /&gt;
[https://www.yorkregion.com/news/power-company-executive-responds-to-york-region-outages/article_c4d072e7-2892-5c9c-8deb-ac5e1936779c.html 2]).&lt;br /&gt;
&lt;br /&gt;
'''Thu Feb 19, 2026, 3:00 pm:''' Systems restored. Please report issues to support@scinet.utoronto.ca.&lt;br /&gt;
&lt;br /&gt;
'''Tue Feb 17, 2026, 8:40 am:''' Power outage at the data centre.  Cooling issues have developed as a result.  Major systems (Trillium, S4H) are expected to be down until sometime Thursday. Login nodes and file systems will remain accessible.&lt;br /&gt;
&lt;br /&gt;
'''Mon Feb 16, 2026, 8:40 pm:''' Electricity is unstable in the data centre area due to severe snowfall.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 1:40 pm:''' All services are operational again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 12:00 pm:''' The Trillium and Open OnDemand compute nodes are operational again. We are still working on bringing Balam, Neptune and S4H nodes up again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 10:00 am:''' There was a power glitch at the data centre overnight. The login nodes are accessible but the compute nodes are down.  &lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--  When removing system status entries, please archive them to: --&amp;gt;&lt;br /&gt;
[[Previous messages]]&lt;br /&gt;
&lt;br /&gt;
{|style=&amp;quot;border-spacing: 10px;width: 100%&amp;quot;&lt;br /&gt;
|valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== QuickStart Guides ==&lt;br /&gt;
* [https://docs.alliancecan.ca/wiki/Trillium_Quickstart Trillium Quickstart]&lt;br /&gt;
* [[HPSS | HPSS archival storage]]&lt;br /&gt;
* [[Teach|Teach cluster]]&lt;br /&gt;
* [[FAQ | FAQ (frequently asked questions)]]&lt;br /&gt;
* [[Acknowledging SciNet]]&lt;br /&gt;
| valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== Tutorials, Manuals, etc. ==&lt;br /&gt;
* [https://education.scinet.utoronto.ca SciNet education material]&lt;br /&gt;
* [https://www.youtube.com/c/SciNetHPCattheUniversityofToronto SciNet's YouTube channel]&lt;br /&gt;
* [[Modules for Mist]] &lt;br /&gt;
* [[Commercial software]]&lt;br /&gt;
* [[Burst Buffer]]&lt;br /&gt;
* [[SSH#SSH Keys|SSH keys]]&lt;br /&gt;
* [[SSH Tunneling]]&lt;br /&gt;
* [[Visualization]]&lt;br /&gt;
* [[Running Serial Jobs on Niagara]]&lt;br /&gt;
* [[Jupyter Hub]]&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Rzon</name></author>
	</entry>
	<entry>
		<id>https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7586</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7586"/>
		<updated>2026-03-16T18:01:22Z</updated>

		<summary type="html">&lt;p&gt;Rzon: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;__NOTOC__&lt;br /&gt;
{| style=&amp;quot;border-spacing:10px; width: 95%&amp;quot;&lt;br /&gt;
| style=&amp;quot;padding:1em; padding-top:.1em; border:2px solid #0645ad; background-color:#f6f6f6; border-radius:7px&amp;quot;|&lt;br /&gt;
&lt;br /&gt;
==System Status==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- Use &amp;quot;Up&amp;quot;, &amp;quot;Partial&amp;quot; or &amp;quot;Down&amp;quot;; these are templates. --&amp;gt;&lt;br /&gt;
{|style=&amp;quot;width:100%&amp;quot; &lt;br /&gt;
|{{Partial3 | Trillium|https://docs.alliancecan.ca/wiki/Trillium_Quickstart}}&lt;br /&gt;
|{{Up3 | OnDemand|https://docs.alliancecan.ca/wiki/Trillium_Open_OnDemand_Quickstart}}&lt;br /&gt;
|{{Up | Globus |Globus}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Up | HPSS|HPSS}}&lt;br /&gt;
|{{Up | Balam|Balam}}&lt;br /&gt;
|{{Up | S4H | S4H}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Up | Teach|Teach}}&lt;br /&gt;
|{{Up3 | File system|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Storage}}&lt;br /&gt;
|{{Up3 | External Network|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Logging_in}} &lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Mon Mar 16, 2026. 13:30pm''' Recovering slowly.  Most compute Trillium compute nodes are up again. Please resubmit your jobs that crashed. &lt;br /&gt;
&lt;br /&gt;
'''Mon Mar 16, 2026. 12:00pm''' Power glitch at the data centre caused compute nodes to go down.&lt;br /&gt;
&lt;br /&gt;
'''Thu Mar 12, 2026, 4:15 pm''' Connection to Trillium are operational again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Mar 12, 2026, 1:00 pm''' We've had some login issues particularly for Trillium-GPU. We're investigating.&lt;br /&gt;
&lt;br /&gt;
'''Downtime Announcement:'''  The winter cooling tower maintenance for the SciNet data centre will take place on March 24 and 25, 2026, starting at 7:00 a.m. on the 24th.  All SciNet systems (Trillium, OnDemand, Balam, S4H, Teach, as well as hosted equipment) will have their compute nodes shut down. Login nodes, file systems, and the HPSS system will remain available, and&lt;br /&gt;
jobs will be held in the queue until maintenance is complete.  Starting 7am on Mar 23, users are encouraged to submit small and short jobs that may be scheduled before the maintenance begins.&lt;br /&gt;
&lt;br /&gt;
'''Fri Feb 20, 2026, 11:35 pm:''' Power glitch, ~480 compute nodes rebooted. Regional power quality has been quite poor lately ([https://www.yorkregion.com/news/road-salt-blamed-for-power-outages/article_1a36d25d-5f97-56ee-a0c7-c49c7b732d38.html 1],&lt;br /&gt;
[https://www.yorkregion.com/news/power-company-executive-responds-to-york-region-outages/article_c4d072e7-2892-5c9c-8deb-ac5e1936779c.html 2]).&lt;br /&gt;
&lt;br /&gt;
'''Thu Feb 19, 2026, 3:00 pm:''' Systems restored. Please report issues to support@scinet.utoronto.ca.&lt;br /&gt;
&lt;br /&gt;
'''Tue Feb 17, 2026, 8:40 am:''' Power outage at the data centre.  Cooling issues have developed as a result.  Major systems (Trillium, S4H) are expected to be down until sometime Thursday. Login nodes and file systems will remain accessible.&lt;br /&gt;
&lt;br /&gt;
'''Mon Feb 16, 2026, 8:40 pm:''' Electricity is unstable in the data centre area due to severe snowfall.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 1:40 pm:''' All services are operational again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 12:00 pm:''' The Trillium and Open OnDemand compute nodes are operational again. We are still working on bringing Balam, Neptune and S4H nodes up again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 10:00 am:''' There was a power glitch at the data centre overnight. The login nodes are accessible but the compute nodes are down.  &lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--  When removing system status entries, please archive them to: --&amp;gt;&lt;br /&gt;
[[Previous messages]]&lt;br /&gt;
&lt;br /&gt;
{|style=&amp;quot;border-spacing: 10px;width: 100%&amp;quot;&lt;br /&gt;
|valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== QuickStart Guides ==&lt;br /&gt;
* [https://docs.alliancecan.ca/wiki/Trillium_Quickstart Trillium Quickstart]&lt;br /&gt;
* [[HPSS | HPSS archival storage]]&lt;br /&gt;
* [[Teach|Teach cluster]]&lt;br /&gt;
* [[FAQ | FAQ (frequently asked questions)]]&lt;br /&gt;
* [[Acknowledging SciNet]]&lt;br /&gt;
| valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== Tutorials, Manuals, etc. ==&lt;br /&gt;
* [https://education.scinet.utoronto.ca SciNet education material]&lt;br /&gt;
* [https://www.youtube.com/c/SciNetHPCattheUniversityofToronto SciNet's YouTube channel]&lt;br /&gt;
* [[Modules for Mist]] &lt;br /&gt;
* [[Commercial software]]&lt;br /&gt;
* [[Burst Buffer]]&lt;br /&gt;
* [[SSH#SSH Keys|SSH keys]]&lt;br /&gt;
* [[SSH Tunneling]]&lt;br /&gt;
* [[Visualization]]&lt;br /&gt;
* [[Running Serial Jobs on Niagara]]&lt;br /&gt;
* [[Jupyter Hub]]&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Rzon</name></author>
	</entry>
	<entry>
		<id>https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7583</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7583"/>
		<updated>2026-03-16T16:57:54Z</updated>

		<summary type="html">&lt;p&gt;Rzon: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;__NOTOC__&lt;br /&gt;
{| style=&amp;quot;border-spacing:10px; width: 95%&amp;quot;&lt;br /&gt;
| style=&amp;quot;padding:1em; padding-top:.1em; border:2px solid #0645ad; background-color:#f6f6f6; border-radius:7px&amp;quot;|&lt;br /&gt;
&lt;br /&gt;
==System Status==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- Use &amp;quot;Up&amp;quot;, &amp;quot;Partial&amp;quot; or &amp;quot;Down&amp;quot;; these are templates. --&amp;gt;&lt;br /&gt;
{|style=&amp;quot;width:100%&amp;quot; &lt;br /&gt;
|{{Down3 | Trillium|https://docs.alliancecan.ca/wiki/Trillium_Quickstart}}&lt;br /&gt;
|{{Down3 | OnDemand|https://docs.alliancecan.ca/wiki/Trillium_Open_OnDemand_Quickstart}}&lt;br /&gt;
|{{Up | Globus |Globus}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Up | HPSS|HPSS}}&lt;br /&gt;
|{{Down | Balam|Balam}}&lt;br /&gt;
|{{Down | S4H | S4H}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Down | Teach|Teach}}&lt;br /&gt;
|{{Up3 | File system|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Storage}}&lt;br /&gt;
|{{Up3 | External Network|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Logging_in}} &lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Mon Mar 16, 2026. 12:00pm''' Power glitch at the data centre caused compute nodes to go down.&lt;br /&gt;
&lt;br /&gt;
'''Thu Mar 12, 2026, 4:15 pm''' Connection to Trillium are operational again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Mar 12, 2026, 1:00 pm''' We've had some login issues particularly for Trillium-GPU. We're investigating.&lt;br /&gt;
&lt;br /&gt;
'''Downtime Announcement:'''  The winter cooling tower maintenance for the SciNet data centre will take place on March 24 and 25, 2026, starting at 7:00 a.m. on the 24th.  All SciNet systems (Trillium, OnDemand, Balam, S4H, Teach, as well as hosted equipment) will have their compute nodes shut down. Login nodes, file systems, and the HPSS system will remain available, and&lt;br /&gt;
jobs will be held in the queue until maintenance is complete.  Starting 7am on Mar 23, users are encouraged to submit small and short jobs that may be scheduled before the maintenance begins.&lt;br /&gt;
&lt;br /&gt;
'''Fri Feb 20, 2026, 11:35 pm:''' Power glitch, ~480 compute nodes rebooted. Regional power quality has been quite poor lately ([https://www.yorkregion.com/news/road-salt-blamed-for-power-outages/article_1a36d25d-5f97-56ee-a0c7-c49c7b732d38.html 1],&lt;br /&gt;
[https://www.yorkregion.com/news/power-company-executive-responds-to-york-region-outages/article_c4d072e7-2892-5c9c-8deb-ac5e1936779c.html 2]).&lt;br /&gt;
&lt;br /&gt;
'''Thu Feb 19, 2026, 3:00 pm:''' Systems restored. Please report issues to support@scinet.utoronto.ca.&lt;br /&gt;
&lt;br /&gt;
'''Tue Feb 17, 2026, 8:40 am:''' Power outage at the data centre.  Cooling issues have developed as a result.  Major systems (Trillium, S4H) are expected to be down until sometime Thursday. Login nodes and file systems will remain accessible.&lt;br /&gt;
&lt;br /&gt;
'''Mon Feb 16, 2026, 8:40 pm:''' Electricity is unstable in the data centre area due to severe snowfall.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 1:40 pm:''' All services are operational again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 12:00 pm:''' The Trillium and Open OnDemand compute nodes are operational again. We are still working on bringing Balam, Neptune and S4H nodes up again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 10:00 am:''' There was a power glitch at the data centre overnight. The login nodes are accessible but the compute nodes are down.  &lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--  When removing system status entries, please archive them to: --&amp;gt;&lt;br /&gt;
[[Previous messages]]&lt;br /&gt;
&lt;br /&gt;
{|style=&amp;quot;border-spacing: 10px;width: 100%&amp;quot;&lt;br /&gt;
|valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== QuickStart Guides ==&lt;br /&gt;
* [https://docs.alliancecan.ca/wiki/Trillium_Quickstart Trillium Quickstart]&lt;br /&gt;
* [[HPSS | HPSS archival storage]]&lt;br /&gt;
* [[Teach|Teach cluster]]&lt;br /&gt;
* [[FAQ | FAQ (frequently asked questions)]]&lt;br /&gt;
* [[Acknowledging SciNet]]&lt;br /&gt;
| valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== Tutorials, Manuals, etc. ==&lt;br /&gt;
* [https://education.scinet.utoronto.ca SciNet education material]&lt;br /&gt;
* [https://www.youtube.com/c/SciNetHPCattheUniversityofToronto SciNet's YouTube channel]&lt;br /&gt;
* [[Modules for Mist]] &lt;br /&gt;
* [[Commercial software]]&lt;br /&gt;
* [[Burst Buffer]]&lt;br /&gt;
* [[SSH#SSH Keys|SSH keys]]&lt;br /&gt;
* [[SSH Tunneling]]&lt;br /&gt;
* [[Visualization]]&lt;br /&gt;
* [[Running Serial Jobs on Niagara]]&lt;br /&gt;
* [[Jupyter Hub]]&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Rzon</name></author>
	</entry>
	<entry>
		<id>https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7580</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7580"/>
		<updated>2026-03-12T20:40:27Z</updated>

		<summary type="html">&lt;p&gt;Rzon: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;__NOTOC__&lt;br /&gt;
{| style=&amp;quot;border-spacing:10px; width: 95%&amp;quot;&lt;br /&gt;
| style=&amp;quot;padding:1em; padding-top:.1em; border:2px solid #0645ad; background-color:#f6f6f6; border-radius:7px&amp;quot;|&lt;br /&gt;
&lt;br /&gt;
==System Status==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- Use &amp;quot;Up&amp;quot;, &amp;quot;Partial&amp;quot; or &amp;quot;Down&amp;quot;; these are templates. --&amp;gt;&lt;br /&gt;
{|style=&amp;quot;width:100%&amp;quot; &lt;br /&gt;
|{{Up3 | Trillium|https://docs.alliancecan.ca/wiki/Trillium_Quickstart}}&lt;br /&gt;
|{{Up3 | OnDemand|https://docs.alliancecan.ca/wiki/Trillium_Open_OnDemand_Quickstart}}&lt;br /&gt;
|{{Up | Globus |Globus}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Up | HPSS|HPSS}}&lt;br /&gt;
|{{Up | Balam|Balam}}&lt;br /&gt;
|{{Up | S4H | S4H}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Up | Teach|Teach}}&lt;br /&gt;
|{{Up3 | File system|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Storage}}&lt;br /&gt;
|{{Up3 | External Network|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Logging_in}} &lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Thu Mar 12, 2026, 4:15 pm''' Connection to Trillium are operational again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Mar 12, 2026, 1:00 pm''' We've had some login issues particularly for Trillium-GPU. We're investigating.&lt;br /&gt;
&lt;br /&gt;
'''Downtime Announcement:'''  The winter cooling tower maintenance for the SciNet data centre will take place on March 24 and 25, 2026, starting at 7:00 a.m. on the 24th.  All SciNet systems (Trillium, OnDemand, Balam, S4H, Teach, as well as hosted equipment) will have their compute nodes shut down. Login nodes, file systems, and the HPSS system will remain available, and&lt;br /&gt;
jobs will be held in the queue until maintenance is complete.  Starting 7am on Mar 23, users are encouraged to submit small and short jobs that may be scheduled before the maintenance begins.&lt;br /&gt;
&lt;br /&gt;
'''Fri Feb 20, 2026, 11:35 pm:''' Power glitch, ~480 compute nodes rebooted. Regional power quality has been quite poor lately ([https://www.yorkregion.com/news/road-salt-blamed-for-power-outages/article_1a36d25d-5f97-56ee-a0c7-c49c7b732d38.html 1],&lt;br /&gt;
[https://www.yorkregion.com/news/power-company-executive-responds-to-york-region-outages/article_c4d072e7-2892-5c9c-8deb-ac5e1936779c.html 2]).&lt;br /&gt;
&lt;br /&gt;
'''Thu Feb 19, 2026, 3:00 pm:''' Systems restored. Please report issues to support@scinet.utoronto.ca.&lt;br /&gt;
&lt;br /&gt;
'''Tue Feb 17, 2026, 8:40 am:''' Power outage at the data centre.  Cooling issues have developed as a result.  Major systems (Trillium, S4H) are expected to be down until sometime Thursday. Login nodes and file systems will remain accessible.&lt;br /&gt;
&lt;br /&gt;
'''Mon Feb 16, 2026, 8:40 pm:''' Electricity is unstable in the data centre area due to severe snowfall.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 1:40 pm:''' All services are operational again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 12:00 pm:''' The Trillium and Open OnDemand compute nodes are operational again. We are still working on bringing Balam, Neptune and S4H nodes up again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 10:00 am:''' There was a power glitch at the data centre overnight. The login nodes are accessible but the compute nodes are down.  &lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--  When removing system status entries, please archive them to: --&amp;gt;&lt;br /&gt;
[[Previous messages]]&lt;br /&gt;
&lt;br /&gt;
{|style=&amp;quot;border-spacing: 10px;width: 100%&amp;quot;&lt;br /&gt;
|valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== QuickStart Guides ==&lt;br /&gt;
* [https://docs.alliancecan.ca/wiki/Trillium_Quickstart Trillium Quickstart]&lt;br /&gt;
* [[HPSS | HPSS archival storage]]&lt;br /&gt;
* [[Teach|Teach cluster]]&lt;br /&gt;
* [[FAQ | FAQ (frequently asked questions)]]&lt;br /&gt;
* [[Acknowledging SciNet]]&lt;br /&gt;
| valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== Tutorials, Manuals, etc. ==&lt;br /&gt;
* [https://education.scinet.utoronto.ca SciNet education material]&lt;br /&gt;
* [https://www.youtube.com/c/SciNetHPCattheUniversityofToronto SciNet's YouTube channel]&lt;br /&gt;
* [[Modules for Mist]] &lt;br /&gt;
* [[Commercial software]]&lt;br /&gt;
* [[Burst Buffer]]&lt;br /&gt;
* [[SSH#SSH Keys|SSH keys]]&lt;br /&gt;
* [[SSH Tunneling]]&lt;br /&gt;
* [[Visualization]]&lt;br /&gt;
* [[Running Serial Jobs on Niagara]]&lt;br /&gt;
* [[Jupyter Hub]]&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Rzon</name></author>
	</entry>
	<entry>
		<id>https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7577</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7577"/>
		<updated>2026-03-12T19:21:34Z</updated>

		<summary type="html">&lt;p&gt;Rzon: /* System Status */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;__NOTOC__&lt;br /&gt;
{| style=&amp;quot;border-spacing:10px; width: 95%&amp;quot;&lt;br /&gt;
| style=&amp;quot;padding:1em; padding-top:.1em; border:2px solid #0645ad; background-color:#f6f6f6; border-radius:7px&amp;quot;|&lt;br /&gt;
&lt;br /&gt;
==System Status==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- Use &amp;quot;Up&amp;quot;, &amp;quot;Partial&amp;quot; or &amp;quot;Down&amp;quot;; these are templates. --&amp;gt;&lt;br /&gt;
{|style=&amp;quot;width:100%&amp;quot; &lt;br /&gt;
|{{Up3 | Trillium|https://docs.alliancecan.ca/wiki/Trillium_Quickstart}}&lt;br /&gt;
|{{Up3 | OnDemand|https://docs.alliancecan.ca/wiki/Trillium_Open_OnDemand_Quickstart}}&lt;br /&gt;
|{{Up | Globus |Globus}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Up | HPSS|HPSS}}&lt;br /&gt;
|{{Up | Balam|Balam}}&lt;br /&gt;
|{{Up | S4H | S4H}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Up | Teach|Teach}}&lt;br /&gt;
|{{Up3 | File system|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Storage}}&lt;br /&gt;
|{{Partial3 | External Network|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Logging_in}} &lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Thu Mar 12, 2026, 1:00 pm''' We've had some login issues particularly for Trillium-GPU. We're investigating.&lt;br /&gt;
&lt;br /&gt;
'''Downtime Announcement:'''  The winter cooling tower maintenance for the SciNet data centre will take place on March 24 and 25, 2026, starting at 7:00 a.m. on the 24th.  All SciNet systems (Trillium, OnDemand, Balam, S4H, Teach, as well as hosted equipment) will have their compute nodes shut down. Login nodes, file systems, and the HPSS system will remain available, and&lt;br /&gt;
jobs will be held in the queue until maintenance is complete.  Starting 7am on Mar 23, users are encouraged to submit small and short jobs that may be scheduled before the maintenance begins.&lt;br /&gt;
&lt;br /&gt;
'''Fri Feb 20, 2026, 11:35 pm:''' Power glitch, ~480 compute nodes rebooted. Regional power quality has been quite poor lately ([https://www.yorkregion.com/news/road-salt-blamed-for-power-outages/article_1a36d25d-5f97-56ee-a0c7-c49c7b732d38.html 1],&lt;br /&gt;
[https://www.yorkregion.com/news/power-company-executive-responds-to-york-region-outages/article_c4d072e7-2892-5c9c-8deb-ac5e1936779c.html 2]).&lt;br /&gt;
&lt;br /&gt;
'''Thu Feb 19, 2026, 3:00 pm:''' Systems restored. Please report issues to support@scinet.utoronto.ca.&lt;br /&gt;
&lt;br /&gt;
'''Tue Feb 17, 2026, 8:40 am:''' Power outage at the data centre.  Cooling issues have developed as a result.  Major systems (Trillium, S4H) are expected to be down until sometime Thursday. Login nodes and file systems will remain accessible.&lt;br /&gt;
&lt;br /&gt;
'''Mon Feb 16, 2026, 8:40 pm:''' Electricity is unstable in the data centre area due to severe snowfall.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 1:40 pm:''' All services are operational again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 12:00 pm:''' The Trillium and Open OnDemand compute nodes are operational again. We are still working on bringing Balam, Neptune and S4H nodes up again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 10:00 am:''' There was a power glitch at the data centre overnight. The login nodes are accessible but the compute nodes are down.  &lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--  When removing system status entries, please archive them to: --&amp;gt;&lt;br /&gt;
[[Previous messages]]&lt;br /&gt;
&lt;br /&gt;
{|style=&amp;quot;border-spacing: 10px;width: 100%&amp;quot;&lt;br /&gt;
|valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== QuickStart Guides ==&lt;br /&gt;
* [https://docs.alliancecan.ca/wiki/Trillium_Quickstart Trillium Quickstart]&lt;br /&gt;
* [[HPSS | HPSS archival storage]]&lt;br /&gt;
* [[Teach|Teach cluster]]&lt;br /&gt;
* [[FAQ | FAQ (frequently asked questions)]]&lt;br /&gt;
* [[Acknowledging SciNet]]&lt;br /&gt;
| valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== Tutorials, Manuals, etc. ==&lt;br /&gt;
* [https://education.scinet.utoronto.ca SciNet education material]&lt;br /&gt;
* [https://www.youtube.com/c/SciNetHPCattheUniversityofToronto SciNet's YouTube channel]&lt;br /&gt;
* [[Modules for Mist]] &lt;br /&gt;
* [[Commercial software]]&lt;br /&gt;
* [[Burst Buffer]]&lt;br /&gt;
* [[SSH#SSH Keys|SSH keys]]&lt;br /&gt;
* [[SSH Tunneling]]&lt;br /&gt;
* [[Visualization]]&lt;br /&gt;
* [[Running Serial Jobs on Niagara]]&lt;br /&gt;
* [[Jupyter Hub]]&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Rzon</name></author>
	</entry>
	<entry>
		<id>https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7574</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7574"/>
		<updated>2026-03-06T20:49:02Z</updated>

		<summary type="html">&lt;p&gt;Rzon: /* System Status */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;__NOTOC__&lt;br /&gt;
{| style=&amp;quot;border-spacing:10px; width: 95%&amp;quot;&lt;br /&gt;
| style=&amp;quot;padding:1em; padding-top:.1em; border:2px solid #0645ad; background-color:#f6f6f6; border-radius:7px&amp;quot;|&lt;br /&gt;
&lt;br /&gt;
==System Status==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- Use &amp;quot;Up&amp;quot;, &amp;quot;Partial&amp;quot; or &amp;quot;Down&amp;quot;; these are templates. --&amp;gt;&lt;br /&gt;
{|style=&amp;quot;width:100%&amp;quot; &lt;br /&gt;
|{{Up3 | Trillium|https://docs.alliancecan.ca/wiki/Trillium_Quickstart}}&lt;br /&gt;
|{{Up3 | OnDemand|https://docs.alliancecan.ca/wiki/Trillium_Open_OnDemand_Quickstart}}&lt;br /&gt;
|{{Up | Globus |Globus}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Up | HPSS|HPSS}}&lt;br /&gt;
|{{Up | Balam|Balam}}&lt;br /&gt;
|{{Up | S4H | S4H}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Up | Teach|Teach}}&lt;br /&gt;
|{{Up3 | File system|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Storage}}&lt;br /&gt;
|{{Up3 | External Network|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Logging_in}} &lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Downtime Announcement:'''  The winter cooling tower maintenance for the SciNet data centre will take place on March 24 and 25, 2026, starting at 7:00 a.m. on the 24th.  All SciNet systems (Trillium, OnDemand, Balam, S4H, Teach, as well as hosted equipment) will have their compute nodes shut down. Login nodes, file systems, and the HPSS system will remain available, and&lt;br /&gt;
jobs will be held in the queue until maintenance is complete.  Starting 7am on Mar 23, users are encouraged to submit small and short jobs that may be scheduled before the maintenance begins.&lt;br /&gt;
&lt;br /&gt;
'''Fri Feb 20, 2026, 11:35 pm:''' Power glitch, ~480 compute nodes rebooted. Regional power quality has been quite poor lately ([https://www.yorkregion.com/news/road-salt-blamed-for-power-outages/article_1a36d25d-5f97-56ee-a0c7-c49c7b732d38.html 1],&lt;br /&gt;
[https://www.yorkregion.com/news/power-company-executive-responds-to-york-region-outages/article_c4d072e7-2892-5c9c-8deb-ac5e1936779c.html 2]).&lt;br /&gt;
&lt;br /&gt;
'''Thu Feb 19, 2026, 3:00 pm:''' Systems restored. Please report issues to support@scinet.utoronto.ca.&lt;br /&gt;
&lt;br /&gt;
'''Tue Feb 17, 2026, 8:40 am:''' Power outage at the data centre.  Cooling issues have developed as a result.  Major systems (Trillium, S4H) are expected to be down until sometime Thursday. Login nodes and file systems will remain accessible.&lt;br /&gt;
&lt;br /&gt;
'''Mon Feb 16, 2026, 8:40 pm:''' Electricity is unstable in the data centre area due to severe snowfall.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 1:40 pm:''' All services are operational again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 12:00 pm:''' The Trillium and Open OnDemand compute nodes are operational again. We are still working on bringing Balam, Neptune and S4H nodes up again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 10:00 am:''' There was a power glitch at the data centre overnight. The login nodes are accessible but the compute nodes are down.  &lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--  When removing system status entries, please archive them to: --&amp;gt;&lt;br /&gt;
[[Previous messages]]&lt;br /&gt;
&lt;br /&gt;
{|style=&amp;quot;border-spacing: 10px;width: 100%&amp;quot;&lt;br /&gt;
|valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== QuickStart Guides ==&lt;br /&gt;
* [https://docs.alliancecan.ca/wiki/Trillium_Quickstart Trillium Quickstart]&lt;br /&gt;
* [[HPSS | HPSS archival storage]]&lt;br /&gt;
* [[Teach|Teach cluster]]&lt;br /&gt;
* [[FAQ | FAQ (frequently asked questions)]]&lt;br /&gt;
* [[Acknowledging SciNet]]&lt;br /&gt;
| valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== Tutorials, Manuals, etc. ==&lt;br /&gt;
* [https://education.scinet.utoronto.ca SciNet education material]&lt;br /&gt;
* [https://www.youtube.com/c/SciNetHPCattheUniversityofToronto SciNet's YouTube channel]&lt;br /&gt;
* [[Modules for Mist]] &lt;br /&gt;
* [[Commercial software]]&lt;br /&gt;
* [[Burst Buffer]]&lt;br /&gt;
* [[SSH#SSH Keys|SSH keys]]&lt;br /&gt;
* [[SSH Tunneling]]&lt;br /&gt;
* [[Visualization]]&lt;br /&gt;
* [[Running Serial Jobs on Niagara]]&lt;br /&gt;
* [[Jupyter Hub]]&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Rzon</name></author>
	</entry>
	<entry>
		<id>https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7568</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7568"/>
		<updated>2026-03-06T15:29:14Z</updated>

		<summary type="html">&lt;p&gt;Rzon: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;__NOTOC__&lt;br /&gt;
{| style=&amp;quot;border-spacing:10px; width: 95%&amp;quot;&lt;br /&gt;
| style=&amp;quot;padding:1em; padding-top:.1em; border:2px solid #0645ad; background-color:#f6f6f6; border-radius:7px&amp;quot;|&lt;br /&gt;
&lt;br /&gt;
==System Status==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- Use &amp;quot;Up&amp;quot;, &amp;quot;Partial&amp;quot; or &amp;quot;Down&amp;quot;; these are templates. --&amp;gt;&lt;br /&gt;
{|style=&amp;quot;width:100%&amp;quot; &lt;br /&gt;
|{{Up3 | Trillium|https://docs.alliancecan.ca/wiki/Trillium_Quickstart}}&lt;br /&gt;
|{{Up3 | OnDemand|https://docs.alliancecan.ca/wiki/Trillium_Open_OnDemand_Quickstart}}&lt;br /&gt;
|{{Up | Globus |Globus}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Up | HPSS|HPSS}}&lt;br /&gt;
|{{Up | Balam|Balam}}&lt;br /&gt;
|{{Up | S4H | S4H}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Up | Teach|Teach}}&lt;br /&gt;
|{{Up3 | File system|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Storage}}&lt;br /&gt;
|{{Up3 | External Network|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Logging_in}} &lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Fri Feb 20, 2026, 11:35 pm:''' Power glitch, ~480 compute nodes rebooted. Regional power quality has been quite poor lately ([https://www.yorkregion.com/news/road-salt-blamed-for-power-outages/article_1a36d25d-5f97-56ee-a0c7-c49c7b732d38.html 1],&lt;br /&gt;
[https://www.yorkregion.com/news/power-company-executive-responds-to-york-region-outages/article_c4d072e7-2892-5c9c-8deb-ac5e1936779c.html 2]).&lt;br /&gt;
&lt;br /&gt;
'''Thu Feb 19, 2026, 3:00 pm:''' Systems restored. Please report issues to support@scinet.utoronto.ca.&lt;br /&gt;
&lt;br /&gt;
'''Tue Feb 17, 2026, 8:40 am:''' Power outage at the data centre.  Cooling issues have developed as a result.  Major systems (Trillium, S4H) are expected to be down until sometime Thursday. Login nodes and file systems will remain accessible.&lt;br /&gt;
&lt;br /&gt;
'''Mon Feb 16, 2026, 8:40 pm:''' Electricity is unstable in the data centre area due to severe snowfall.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 1:40 pm:''' All services are operational again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 12:00 pm:''' The Trillium and Open OnDemand compute nodes are operational again. We are still working on bringing Balam, Neptune and S4H nodes up again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 10:00 am:''' There was a power glitch at the data centre overnight. The login nodes are accessible but the compute nodes are down.  &lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--  When removing system status entries, please archive them to: --&amp;gt;&lt;br /&gt;
[[Previous messages]]&lt;br /&gt;
&lt;br /&gt;
{|style=&amp;quot;border-spacing: 10px;width: 100%&amp;quot;&lt;br /&gt;
|valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== QuickStart Guides ==&lt;br /&gt;
* [https://docs.alliancecan.ca/wiki/Trillium_Quickstart Trillium Quickstart]&lt;br /&gt;
* [[Niagara Quickstart]]&lt;br /&gt;
* [[HPSS | HPSS archival storage]]&lt;br /&gt;
* [[Teach|Teach cluster]]&lt;br /&gt;
* [[FAQ | FAQ (frequently asked questions)]]&lt;br /&gt;
* [[Acknowledging SciNet]]&lt;br /&gt;
| valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== Tutorials, Manuals, etc. ==&lt;br /&gt;
* [https://education.scinet.utoronto.ca SciNet education material]&lt;br /&gt;
* [https://www.youtube.com/c/SciNetHPCattheUniversityofToronto SciNet's YouTube channel]&lt;br /&gt;
* [[Modules specific to Niagara|Software Modules specific to Niagara]] &lt;br /&gt;
* [[Modules for Mist]] &lt;br /&gt;
* [[Commercial software]]&lt;br /&gt;
* [[Burst Buffer]]&lt;br /&gt;
* [[SSH#SSH Keys|SSH keys]]&lt;br /&gt;
* [[SSH Tunneling]]&lt;br /&gt;
* [[Visualization]]&lt;br /&gt;
* [[Running Serial Jobs on Niagara]]&lt;br /&gt;
* [[Jupyter Hub]]&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Rzon</name></author>
	</entry>
	<entry>
		<id>https://docs.scinet.utoronto.ca/index.php?title=Previous_messages&amp;diff=7565</id>
		<title>Previous messages</title>
		<link rel="alternate" type="text/html" href="https://docs.scinet.utoronto.ca/index.php?title=Previous_messages&amp;diff=7565"/>
		<updated>2026-03-06T15:29:04Z</updated>

		<summary type="html">&lt;p&gt;Rzon: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;'''Thu Jan 16, 2026, 11:00 pm:''' HPSS is back online, and accessible via alliancecan#hpss Globus endpoint. &lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 15, 2026, 10:00 pm:''' HPSS will undergo maintenance on Friday morning, Jan/16/2025, , including alliancecan#hpss Globus endpoint &lt;br /&gt;
&lt;br /&gt;
'''Tue Jan 6, 2026, 10:15 am:''' OnDemand has been fixed and is working again.&lt;br /&gt;
&lt;br /&gt;
'''Mon Jan 5, 2026, 9:00 pm:''' The authentication mechanism of OnDemand is not working.&lt;br /&gt;
&lt;br /&gt;
'''Wed Dec 31, 2025, 12:40 pm:''' We believe the problem has now been resolved.  Please let us know if you still experience login problems or aborted jobs.&lt;br /&gt;
&lt;br /&gt;
'''Tue Dec 30, 2025, 2:10 pm:''' We are experiencing problems with authentication, resulting in failed logins, OOD errors, and aborted jobs (with &amp;quot;prolog error&amp;quot;).  Please bear with us, as we are very short-staffed during the holiday break.  We will post updates here.&lt;br /&gt;
&lt;br /&gt;
'''Tue Dec 3, 2025, 11:30 am:''' Open OnDemand is fully operational again.&lt;br /&gt;
&lt;br /&gt;
'''Sat Nov 29, 2025, 00:40 am:''' There has been a problem with the water chiller. Some systems are offline.&lt;br /&gt;
&lt;br /&gt;
'''Wed Nov 5, 2025, 12:55 pm:''' Balam is back online.&lt;br /&gt;
&lt;br /&gt;
'''Wed Nov 5, 2025, 10:00 am:''' Open OnDemand is back online.&lt;br /&gt;
&lt;br /&gt;
'''Tue Nov 4, 2025, 11:00 pm:''' Most of the work is done, data movers, Globus, and HPSS are back online. Remaining services will be worked on tomorrow.&lt;br /&gt;
&lt;br /&gt;
'''Tue Nov 4, 2025, 8:30 am:''' Scheduled network maintenance. Trillium cluster is *not* affected.&lt;br /&gt;
&lt;br /&gt;
'''Tue Oct 21, 2025, 17:30 am:''' Balam maintenance finished.&lt;br /&gt;
&lt;br /&gt;
'''Tue Oct 21, 2025, 7:00 am:''' Balam maintenance day.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 15, 2025, 3:55 pm:''' Trillium inbound connections through trillium.alliancecan.ca or trillium.scinet.utoronto.ca are working again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 15, 2025, 3:05 pm:''' Trillium is experiencing external network issues for both incoming traffic. Please try: ssh USERNAME@tri-login01.scinet.utoronto.ca in the meantime.&lt;br /&gt;
 &lt;br /&gt;
'''Thu Oct 06, 2025, 8:00 pm:''' HPSS is fully functional. You may submit archive jobs from trillium login nodes, datamovers and robots.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 03, 2025, 6:30 pm:''' HPSS is back online, and already accessible via alliancecan#hpss Globus endpoint. Directory tree now follows the other Alliance clusters. We're still working on job submission via Slurm&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 01, 2025, 0:00 am:''' Niagara compute nodes are now unavailable for regular users. The login nodes will remain available for a while to allow a few last data transfers, although transfers from the Niagara file systems to Trillium are best done on nia-dm1.scinet.utoronto.ca.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 01, 2025, 9:30 am:''' HPSS is down for scheduled maintenance, including alliancecan#hpss Globus endpoint&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
'''Thu Sep 18, 2025, 11:30 am:''' Open OnDemand is fully functional again.&lt;br /&gt;
&lt;br /&gt;
'''Wed Sep 17, 2025, 6:00 pm:''' Niagara is back up as well (including its Globus endpoint).  We are still working on the other systems.&lt;br /&gt;
&lt;br /&gt;
'''Wed Sep 17, 2025, 1:40 pm:''' Trillium is back up (except for its Globus endpoint).  We are working on the other systems still.&lt;br /&gt;
&lt;br /&gt;
'''Tue Sep 16, 2025, 5:45 pm:''' Unfortunately, we cannot bring all systems up yet because we are waiting for a spare part for the cooling system that will be brought tomorrow.  In the meantime, we have managed to keep the Trillium login nodes up, but not other systems.&lt;br /&gt;
&lt;br /&gt;
'''Tue Sep 16, 2025, from 7:00 am to 5:00 pm (EDT):''' The SciNet datacentre will undergo maintenance of several critical parts of the centre.  This will require a full shutdown of all SciNet systems (Trillium, Niagara, Mist, HPSS, Rouge, Teach, as well as hosted equipment). This will also be the time that the Mist cluster gets decommissioned. &lt;br /&gt;
&lt;br /&gt;
'''Fri Sep 12 22:03:17 EDT 2025:''' HPSS software and OS upgrades are finished.&lt;br /&gt;
&lt;br /&gt;
'''Tue Sep  9 17:05:38 EDT 2025:''' Starting tomorrow, Sep/10, and for the following 3 days HPSS will be down for software and OS upgrades. We will strive to finish sooner, at which time we will make the system available to users again.&lt;br /&gt;
&lt;br /&gt;
===Mist/Niagara Decommissioning Schedule===&lt;br /&gt;
&lt;br /&gt;
'''September 4, 2025'''&lt;br /&gt;
* Niagara reduced to 863 compute nodes.&lt;br /&gt;
&lt;br /&gt;
'''September 9, 2025'''&lt;br /&gt;
* Niagara's Open OnDemand decommissioned.&lt;br /&gt;
* Brief data centre connection outage at 9 AM EDT&lt;br /&gt;
* Niagara reduced to 647 compute nodes at end of day.&lt;br /&gt;
&lt;br /&gt;
'''September 11, 2025'''&lt;br /&gt;
* Trillium Open OnDemand goes live.&lt;br /&gt;
&lt;br /&gt;
'''September 16, 2025'''&lt;br /&gt;
* '''Full-day data centre maintenance'''&lt;br /&gt;
* Niagara reduced to 431 compute nodes.&lt;br /&gt;
* Mist decommissioned.&lt;br /&gt;
&lt;br /&gt;
'''September 24, 2025'''&lt;br /&gt;
* Niagara reduced to 215 compute nodes.&lt;br /&gt;
&lt;br /&gt;
'''September 30, 2025'''&lt;br /&gt;
* Niagara decommissioned.&lt;br /&gt;
&lt;br /&gt;
'''August 25, 2025, 9:50 EDT:''' Open Ondemand is fully operational again.&lt;br /&gt;
&lt;br /&gt;
'''August 22, 2025, 3:15 PM EDT:''' Open Ondemand has issues launching new interactive apps. We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''August 20, 2025, 10:00 AM EDT:''' The GPU scheduler on Trillium is fully operational again.&lt;br /&gt;
&lt;br /&gt;
'''August 19, 2025, 5:00 PM EDT:''' The GPU scheduler on Trillium has trouble scheduling multi-GPU jobs.  We're investigating the issue.&lt;br /&gt;
&lt;br /&gt;
'''August 7, 2025:''' CVMFS issues are resolved.&lt;br /&gt;
&lt;br /&gt;
'''August 6, 2025:''' We are seeing intermittent issues with the software on CVMFS on Niagara. We're investigating.&lt;br /&gt;
&lt;br /&gt;
'''July 31, 2025, 4:00 PM EDT - 5:00 PM EDT:''' As announced, all systems connected to the Niagara file system (Mist, Niagara, HPSS, Balam, and Rouge) will be paused and inaccessible for one hour to start the transfer of files from the Niagara file system to the Trillium file system. &lt;br /&gt;
&lt;br /&gt;
'''January 6, 2025:''' As part of the installation of the new computing cluster Trillium, there is now a permanent reduction in computing capacity of Niagara to 50% and of Mist to 35%.&lt;br /&gt;
&lt;br /&gt;
'''July 9, 2025:''' The [[Teach]] cluster will be unavailable for the day for network maintenance.&lt;br /&gt;
&lt;br /&gt;
'''July 4, 2025:''' Open OnDemand is back up.&lt;br /&gt;
&lt;br /&gt;
'''July 4, 2025:''' Open OnDemand is down. We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''June 25, 2025, 7:15 PM EDT:''' The [[Teach]] cluster's scheduler is up again.&lt;br /&gt;
&lt;br /&gt;
'''June 25, 2025, 4:30 PM EDT:''' The [[Teach]] cluster's scheduler is down. We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''April 30, 2025, 9:30 AM EDT:''' The [[Teach]] cluster is available again.&lt;br /&gt;
&lt;br /&gt;
'''April 30, 2025:''' The [[Teach]] cluster will be unavailable from 8:00 am to about 12:00 noon for file system maintenance.&lt;br /&gt;
&lt;br /&gt;
'''April 1, 2025:''' The Jupyter Hub has been replaced by SciNet's [[Open OnDemand Quickstart|Open OnDemand service]].&lt;br /&gt;
&lt;br /&gt;
'''March 1, 2025:''' As of March 1st scratch purging is suspended until after Trillium comes online.&lt;br /&gt;
&lt;br /&gt;
'''April 15, 2025 12:40 pm EDT: '''Balam login node is accessible again.&lt;br /&gt;
&lt;br /&gt;
'''April 15, 2025 12:10 pm EDT: '''Balam login node is under maintenance and temporarily inaccessible to users.&lt;br /&gt;
&lt;br /&gt;
'''April 9, 2025 9PM:''' HPSS is back online.&lt;br /&gt;
&lt;br /&gt;
'''April 8, 2025 9PM:''' HPSS is being reserved for OS updates on April 9 (Wednesday).&lt;br /&gt;
&lt;br /&gt;
'''March 31, 2025 3:20 pm EDT:''' Mist login node is accessible again.&lt;br /&gt;
&lt;br /&gt;
'''March 31, 2025 2:45 pm EDT:''' Mist login node is under maintenance and temporarily inaccessible to users.&lt;br /&gt;
&lt;br /&gt;
'''March 28, 2025 3:00 pm - 4:00 pm EDT:''' A short maintenance was needed for the Teach compute nodes; you might have experienced some job scheduling delays on that cluster. &lt;br /&gt;
&lt;br /&gt;
'''March 20, 2025 10:30 am EDT:''' Teach compute nodes are back. &lt;br /&gt;
&lt;br /&gt;
'''March 19, 2025 11:00 pm EDT:''' Teach compute nodes are down again. &lt;br /&gt;
&lt;br /&gt;
'''March 19, 2025 5:15pm EDT:''' Maintenance of the cooling system was performed successfully. The cluster is back on line&lt;br /&gt;
&lt;br /&gt;
'''March 19, 2025 8:00 am - 5:00 pm EDT:''' Maintenance of the cooling system as well as preparations for the Trillium cluster will require a shutdown of the compute nodes of all SciNet systems (Niagara, Mist, Rouge, Balam, Teach, as well as hosted equipment). The login nodes, file systems and the HPSS system will remain available. The scheduler will hold jobs that are submitted until the maintenance has finished.&lt;br /&gt;
&lt;br /&gt;
'''March 18, 2025 10:00 am EDT:''' Teach compute nodes are back.&lt;br /&gt;
&lt;br /&gt;
'''March 17, 2025 10:00 pm EDT:''' Teach compute nodes are down. We are working on it. &lt;br /&gt;
&lt;br /&gt;
'''February 27, 2025 9:00 pm EST:''' Access to HPSS via Globus has been restored.&lt;br /&gt;
&lt;br /&gt;
'''February 25, 2025 2:30 pm EST:''' Access to HPSS via Globus is currently suspended (sorry, trivial upgrade has gone wrong).&lt;br /&gt;
&lt;br /&gt;
'''February 25, 2025 12:30 pm EST:''' Mist login node is accessible again.&lt;br /&gt;
&lt;br /&gt;
'''February 25, 2025 11:50 am EST:''' Mist login node is under maintenance and temporarily inaccessible to users.&lt;br /&gt;
&lt;br /&gt;
'''February 7, 2025 2:45 pm EST:''' Systems are back online.&lt;br /&gt;
&lt;br /&gt;
'''Fri Feb  7 01:04:33 EST 2025:''' There has been a problem with the water chiller. Automatic thermal shutdown of the compute nodes&lt;br /&gt;
&lt;br /&gt;
'''January 31, 2025 11:45 am EST:''' Power is back.&lt;br /&gt;
&lt;br /&gt;
'''January 31, 2025 6:00 am EST:''' Power outage in the data center. Many compute jobs will have stopped. Until power gets restored, parts of the systems are  running on the generator. No ETA on full power restoration.&lt;br /&gt;
 &lt;br /&gt;
'''January 28, 2025 9:30 pm EST:''' The CCEnv stack has been restored.&lt;br /&gt;
&lt;br /&gt;
'''January 28, 2025 5:00 pm EST:''' The CCEnv stack from cvmfs has issues and may not work reliable.&lt;br /&gt;
&lt;br /&gt;
'''January 23, 2025 9:00 am - 1:00 pm EST:''' Balam, Rouge and Neptune compute nodes will be shut down from 9 AM to 1 PM EST for additional electrical work.&lt;br /&gt;
&lt;br /&gt;
'''January 22, 2025 12:55 pm EST:''' Compute nodes are back online&lt;br /&gt;
&lt;br /&gt;
'''January 22, 2025 8:00 am - 5:00 pm EST:''' Preparations for the new system Trillium will require a shutdown of the compute nodes of all SciNet systems (Niagara, Mist, Rouge, Teach, as well as hosted equipment) from 8 AM to 5 PM EST. The login nodes, file systems and the HPSS system will remain available. The scheduler will hold jobs that are submitted until the maintenance has finished.&lt;br /&gt;
&lt;br /&gt;
'''January 9, 2025 11:00 am EST:''' Systems are back online&lt;br /&gt;
&lt;br /&gt;
'''January 8, 2025 10:34 pm EST:''' We had some sort of thermal event at the datacenter, and the clusters are down. We're still investigating&lt;br /&gt;
&lt;br /&gt;
'''January 8, 2025 08:00 am EST:''' Balam, Rouge and Neptune are shutdown for electrical upgrades&lt;br /&gt;
&lt;br /&gt;
'''January 6, 2025:''' As part of the installation of the new computing cluster Trillium, there will be a (permanent) reduction in computing capacity of Niagara and Mist. Only 50% of Niagara and 35% of Mist will remain active after January 6th.  The reduction will require Mist to be shutdown for a few hours on January 6th. Balam, Rouge and Neptune will be shutdown on Wednesday January 8th for the same reason.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
'''December 20, 2024 09:00 am EST:''' OpenOnDemand service will not be available on Dec 20 from 9 a.m. to 5 p.m. due to scheduled maintenance.&lt;br /&gt;
&lt;br /&gt;
'''December 16, 2024, 08:21 am EST:''' The Niagara scheduler has been restarted.&lt;br /&gt;
  &lt;br /&gt;
'''December 16, 2024, 00:04 am EST:''' The Niagara scheduler has an issue; we are investigating.&lt;br /&gt;
  &lt;br /&gt;
'''Fri Nov 8, 2024, 09:45 AM EST.''' Balam and Rouge schedulers are back online.&lt;br /&gt;
&lt;br /&gt;
'''Thu Nov 7, 2024, 10:30 PM EST.''' Most systems are up, except for the schedulers on Balam and Rouge (but even their login nodes are up), and a few 'neptune' niagara nodes.&lt;br /&gt;
&lt;br /&gt;
'''Thu Nov 7, 2024, 5:30 PM EST:''' Systems are being brought up, but not yet available for users.&lt;br /&gt;
&lt;br /&gt;
'''Downtime Announcement: On Thu Nov 7, 2024, all systems and storage located at the SciNet Datacenter (Niagara, Mist, HPSS, Rouge, Teach, JupyterHub, Balam) will be unavailable from 7 a.m. to 5 p.m. ET.&lt;br /&gt;
This outage is required to install new electrical equipment (UPS) for the upcoming systems refresh. The work is expected to be completed in one day.&lt;br /&gt;
The scheduler will hold jobs that cannot finish before the start of the shutdown. Users are encouraged to submit small and short jobs that can take advantage of this, as the scheduler may be able to fit these jobs in before the maintenance on otherwise idle nodes.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 24 15:05 EDT 2024''': Cooling pump motor has been replaced. All systems are back to normal.&lt;br /&gt;
&lt;br /&gt;
'''Tue Oct 22 16:35 EDT 2024''': The motor is scheduled for replacement on Thursday, Oct 24.&lt;br /&gt;
&lt;br /&gt;
'''Mon Oct 21 17:15 EDT 2024''': Compute nodes will remain down until we can replace the main cooling pump.  This may take several days.  Please see this page for updates.&lt;br /&gt;
&lt;br /&gt;
'''Mon Oct 21 12:15 EDT 2024''': Compute nodes have been shutdown due to a cooling system failure.&lt;br /&gt;
&lt;br /&gt;
'''Fri Oct 18 21:40 EDT 2024''': Systems are back to normal&lt;br /&gt;
&lt;br /&gt;
'''Fri Oct 18 21:15 EDT 2024''': We are experiences technical difficulties, apparently caused by a glitch in the file systems&lt;br /&gt;
&lt;br /&gt;
'''Tue Oct 1 10:45 EDT 2024''': The Jupyter Hub service will be rebooted today at around 11:00 am EDT for system upgrades. &lt;br /&gt;
&lt;br /&gt;
'''Tue Sep 3 07:00 EDT 2024''': Intermittent file system issues which may cause issues logging in.  We are in the process of resolving the issue.&lt;br /&gt;
&lt;br /&gt;
'''Sun Sep 1 00:01 - 04:00 EDT 2024''': Network maintenance may cause connection issues to the datacentre.&lt;br /&gt;
&lt;br /&gt;
'''Thu Aug 22 13:30:00 EDT 2024''': Chiller issue caused about 25% of Niagara compute nodes to go down; users should resubmit any affected jobs.&lt;br /&gt;
&lt;br /&gt;
'''Wed Aug 21 16:35:00 EDT 2024''': Maintenance finished; compute nodes are now available for user jobs.&lt;br /&gt;
&lt;br /&gt;
'''Wed Aug 21 7:00:00 EDT 2024''': Maintenance started.&lt;br /&gt;
&lt;br /&gt;
'''Sun Aug 18 19:15:00 EDT 2024''': Issues have been resolved.&lt;br /&gt;
&lt;br /&gt;
'''Sun Aug 18 14:30:00 EDT 2024''': Power issues seem to have brought compute nodes down, and compounded to the file system issues we had earlier.&lt;br /&gt;
&lt;br /&gt;
'''Sun Aug 18 10:31:53 EDT 2024''': GPFS is back online, and seems to be holding&lt;br /&gt;
&lt;br /&gt;
'''Sun Aug 18 08:44:40 EDT 2024''': Sorry, problems with GPFS file systems are reoccurring. &lt;br /&gt;
&lt;br /&gt;
'''Sun Aug 18 07:59:02 EDT 2024''': GPFS file systems are back to normal. Many jobs have died and will need to be resubmitted.&lt;br /&gt;
&lt;br /&gt;
'''Sun Aug 18 06:39:12 EDT 2024''': Support staff detected the problem and started to work on the fix&lt;br /&gt;
&lt;br /&gt;
'''Sun Aug 18 00:53:52 EDT 2024''': GPFS file systems (home, scratch, project) started to show initial stages of problems&lt;br /&gt;
&lt;br /&gt;
'''August 21, 2024''': The annual cooling tower maintenance for the SciNet data centre will take place on August 21, 2024 from 7 a.m. EDT until the end of day. This maintenance requires a shutdown of the compute nodes of all SciNet systems (Niagara, Mist, Rouge, Teach, as well as hosted equipment). The login nodes, file systems and the HPSS system will remain available.&lt;br /&gt;
&lt;br /&gt;
The scheduler will hold jobs that cannot finish before the start of the shutdown. Users are encouraged to submit small and short jobs that can take advantage of this, as the scheduler may be able to fit these jobs in before the maintenance on otherwise idle nodes.&lt;br /&gt;
&lt;br /&gt;
'''Thursday, August 1, 10:00 PM EDT''' Filesystem problems resolved.&lt;br /&gt;
&lt;br /&gt;
'''Thursday, August 1, 9:30 PM EDT''' Filesystem problems preventing logins to the systems.  Working on it.&lt;br /&gt;
&lt;br /&gt;
'''Monday, July 22, 11:50 AM EDT''' Systems are back to normal&lt;br /&gt;
&lt;br /&gt;
'''Monday, July 22, 10:50 AM EDT''' Cooling problem has been fixed. Systems are coming up&lt;br /&gt;
&lt;br /&gt;
'''Monday, July 22, 10:20 AM EDT''' Compute nodes have been shutdown due to a cooling tower failure.&lt;br /&gt;
&lt;br /&gt;
'''Friday, July 19, 9:30 AM EDT''' CCEnv modules available on all login nodes again.&lt;br /&gt;
&lt;br /&gt;
'''Friday, July 19, 5:00 AM EDT''' Some login nodes do not have the CCEnv modules available.  We are working on a fix.&lt;br /&gt;
&lt;br /&gt;
'''Monday, Jun 3, 12:55 PM EDT''' All systems are recovered now.&lt;br /&gt;
&lt;br /&gt;
'''Monday, Jun 3, 10:50 AM EDT''' The file system issues affect all nodes, so all systems are inaccessible to users at the moment. No time estimate yet for when the systems may be back.&lt;br /&gt;
&lt;br /&gt;
'''Monday, Jun 3, 7:58 AM EDT''' Login issues for Niagara and Mist. There are file system issues as well. Investigating.&lt;br /&gt;
&lt;br /&gt;
'''Sunday, Jun 2, 12:00 PM EDT''' CCEnv modules missing, investigating.&lt;br /&gt;
&lt;br /&gt;
'''Wednesday May 29, 5:50 PM EDT''' Niagara compute nodes are up.  &lt;br /&gt;
&lt;br /&gt;
'''Wednesday May 29, 4:40 PM EDT''' Niagara compute nodes are coming up.  &lt;br /&gt;
&lt;br /&gt;
'''Wednesday May 29, 4 PM EDT''' Niagara login nodes and jupyterhub are up; file system is now accessible.  &lt;br /&gt;
&lt;br /&gt;
'''Wednesday May 29, 2 PM EDT''' Electricians are checking and testing all junction boxes and connectors under the raised floor for safety.  Some systems are expected to be back up later today (storage, login nodes), and compute systems will be powered up as soon as it is deemed safe.&lt;br /&gt;
&lt;br /&gt;
'''Tuesday May 28, 3 PM EDT''' Cleaning crews are at the datacentre, to pump the water and install dryers.  Once the floors are dry, we need to inspect all electrical boxes to ensure safety.  We do not expect to have a fully functional datacentre before Thursday, although we hope to be able to turn on the storage and login nodes sometime tomorrow, if circumstances permit.  Apologies, and thank you for your patience.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
'''Tuesday May 28, 7 AM EDT''' A water mains break outside our datacentre has caused extensive flooding, and all systems have been shut down preventatively. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
'''Friday May 17, 10 PM EDT - Saturday May 18, 2 AM EDT:''' The external network will be unavailable for maintenance. Running and queued jobs on the systems will not be affected.&lt;br /&gt;
&lt;br /&gt;
'''Tuesday May 14, 6:45 PM EDT:''' All systems are recovered now.&lt;br /&gt;
&lt;br /&gt;
'''Tuesday May 14, 5 PM EDT:''' Power loss at the datacentre resulted in loss of cooling.  Systems are being restored.&lt;br /&gt;
&lt;br /&gt;
'''Friday May 3, 10 PM EDT - Saturday May 4, 2 AM EDT:''' The external network will be unavailable for maintenance. Running and queued jobs on the systems will not be affected.&lt;br /&gt;
&lt;br /&gt;
'''Tuesday April 17, 2024: 11:00 ''' The restart of the Niagara login nodes has been completed successfully.&lt;br /&gt;
&lt;br /&gt;
'''Tuesday April 17, 2024: 09:40 ''' Niagara login nodes will be rebooted &lt;br /&gt;
&lt;br /&gt;
'''Tuesday April 16, 2024: 12:45 ''' mist-login01  recovered now&lt;br /&gt;
&lt;br /&gt;
'''Tuesday April 16, 2024: 11:45 ''' mist-login01  will be unavailable due to maintenance from 12:15 to 12:45. Following the completion of maintenance, login access should be restored &lt;br /&gt;
&lt;br /&gt;
'''Monday April 15, 2024: 13:02 ''' Balam-login01 will be unavailable due to maintenance from 13:00 to 13:30. Following the completion of maintenance, login access should be restored and available once more. &lt;br /&gt;
&lt;br /&gt;
'''Monday March 18, 2024: 14:45 ''' File system issue resolved.  Users are advised to check if their running jobs were affected, and if so, to resubmit.&lt;br /&gt;
&lt;br /&gt;
'''Monday March 18, 2024: 13:02 ''' File system issues.  This affects the ability to log in. We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''Monday March 11, 2024: 14:05 ''' All systems are recovered now&lt;br /&gt;
&lt;br /&gt;
'''Monday March 11, 2024:''' There will be an shutdown of the file system at SciNet for an emergency repair. As a consequence, the login nodes and compute nodes of all SciNet clusters using the file system (Niagara, Mist, Balam, Rouge, and Teach) will be down from 11 am EST until later in the afternoon. &lt;br /&gt;
&lt;br /&gt;
'''February 28, 2024, 16:30 PM EDT:''' All systems are recovered now.&lt;br /&gt;
&lt;br /&gt;
'''February 28, 2024, 1:00 PM EDT:''' A loop pump fault caused many compute nodes overheat. If you jobs failed around this time, please resubmit. Once the root cause has been addressed, the cluster will be brought up completely. Please report issues to support@scinet.utoronto.ca.&lt;br /&gt;
&lt;br /&gt;
'''February 22, 2024, 5:45 PM EDT:''' Maintenance finished and system restored. Please report issues to support@scinet.utoronto.ca.&lt;br /&gt;
&lt;br /&gt;
'''February 21, 2024, 7:00 AM EDT:''' Maintenance starting.  Niagara login nodes and the file system are kept up as much as possible, but will be rebooted at some point.&lt;br /&gt;
&lt;br /&gt;
'''February 20, 2024, 3:45 PM EDT:''' Cooling tower has been restored, all systems are in production. &lt;br /&gt;
&lt;br /&gt;
'''February 20, 2024, 1:30 AM EDT:''' Cooling tower malfunction, all compute nodes are shutdown, the root cause will be addressed earliest in the morning.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;span style=&amp;quot;color:red&amp;quot;&amp;gt;&amp;lt;b&amp;gt; February 21 and 22, 2024: SciNet Data Centre Maintenance:&amp;lt;/b&amp;gt;&amp;lt;/span&amp;gt;&amp;lt;br/&amp;gt;&lt;br /&gt;
This annual winter maintenance involves a full data centre shutdown&lt;br /&gt;
starting at 7:00 am EST on Wednesday, February 21st.  None of the&lt;br /&gt;
SciNet systems (Niagara, Mist, Rouge, Teach, the file systems, as&lt;br /&gt;
well as hosted equipment) will be accessible.  All systems should be&lt;br /&gt;
fully available again in the last afternoon of the 22nd.&lt;br /&gt;
&lt;br /&gt;
The scheduler will hold jobs that cannot finish before the start of&lt;br /&gt;
the shutdown. Users are encouraged to submit small and short jobs&lt;br /&gt;
that can take advantage of this, as the scheduler may be able to fit&lt;br /&gt;
these jobs in before the maintenance on otherwise idle nodes.&lt;br /&gt;
&lt;br /&gt;
'''Mon January 29, 08:20    (EST):''' Access to Niagara login nodes restored (it was an internal routing issue).&lt;br /&gt;
&lt;br /&gt;
'''Mon January 29, 07:35    (EST):''' No access to Niagara login nodes.  We are investigating.  Use the Mist login to get access to SciNet systems.&lt;br /&gt;
&lt;br /&gt;
'''Wed January 24, 15:20    (EST):''' maintenance on rouge-login01 &lt;br /&gt;
&lt;br /&gt;
'''Wed January 24, 14:55    (EST):''' Rebooting rouge-login01 &lt;br /&gt;
&lt;br /&gt;
'''Tue January 23, 10:25 am (EST):''' Mist-login01 maintenance done &lt;br /&gt;
&lt;br /&gt;
'''Tue January 23, 10:10 am (EST):''' Rebooting Mist-login01 to deploy new image&lt;br /&gt;
&lt;br /&gt;
'''Tue January 22, 21:00 am (EST):''' HPSS performance for hsi &amp;amp; htar clients is back to normal.&lt;br /&gt;
&lt;br /&gt;
'''Tue January 20, 11:50 am (EST):''' HPSS hsi/htar/VFS jobs will remain on PD state on the queue over the weekend, so we may work on archive02/vfs02 on Monday, and try to improve transfer performance. In the meantime you may use Globus (computecanada#hpss) if your workflow is suitable. &lt;br /&gt;
&lt;br /&gt;
'''Tue January 14, 13:20 am (EST):''' The ongoing HPSS jobs from Friday finished earlier, so we restarted HPSS sooner and released the PD jobs on the queue. &lt;br /&gt;
&lt;br /&gt;
'''Tue January 12, 10:40 am (EST):''' We have applied some tweaks to the HPSS configuration to improve performance, but they won't take effect until we restart the services, which scheduled for Monday morning. If over the weekend we notice that there are no HPSS jobs running on the queue we may restart HPSS sooner. &lt;br /&gt;
&lt;br /&gt;
'''Tue January 09, 9:10 am (EST):''' Remaining cvmfs issues cleared.&lt;br /&gt;
&lt;br /&gt;
'''Tue January 09, 8:00 am (EST):''' We're investigating remaining issues with cvmfs access on login nodes.&lt;br /&gt;
&lt;br /&gt;
'''Mon January 08, 21:50 pm (EST):''' File systems are back to normal. Please resubmit your jobs.  &lt;br /&gt;
&lt;br /&gt;
'''Mon January 08, 9:10 pm (EST):''' We had a severe deadlock, and some disk volumes went down. The file systems are being recovered now. It could take another hour.&lt;br /&gt;
&lt;br /&gt;
'''Mon January 08, 7:20 pm (EST):''' We seem to have a problem with the file system, and are investigating.&lt;br /&gt;
&lt;br /&gt;
'''Tue December 19, 2:45 pm (EST):''' Compute nodes are available again.  &lt;br /&gt;
&lt;br /&gt;
'''Tue December 19, 12:09 pm (EST):''' Maintenance was postponed by one hour. &lt;br /&gt;
&lt;br /&gt;
'''Tue December 19, 12 noon - 1 pm (EST):''' There will be a shutdown of the compute nodes of the Niagara, Mist and Rouge cluster to allow for an emergency repair to the cooling tower.  Login nodes will remain available but no jobs will run during that time.  Updates will be posted on here.&lt;br /&gt;
&lt;br /&gt;
'''Mon Dec  11 11:17:00 EST 2023:''' File systems recovered; Niagara and Mist are operational again.&lt;br /&gt;
&lt;br /&gt;
'''Mon Dec  11 7:51:00 EST 2023:''' Niagara's login nodes are being overwhelmed.  We are investigating. Likely file-system related.&lt;br /&gt;
&lt;br /&gt;
'''Thu Dec  6 10:01:24 EST 2023:''' Niagara's scheduler rebooting for security patches.&lt;br /&gt;
&lt;br /&gt;
'''Wed Dec  6 13:06:46 EST 2023:''' Endpoint computecanada#niagara transition from Globus GCSv4 to GCSv5 is completed. computecanada#niagara-GCSv4 has been deactivated&lt;br /&gt;
&lt;br /&gt;
'''Mon Dec  4 16:35:07 EST 2023:''' Endpoint computecanada#niagara has now been upgraded to Globus GCSv5. The old endpoint is still available as computecanada#niagara-GCSv4 on nia-datamover2, only until Wednesday, at which time we'll disable it as well.&lt;br /&gt;
&lt;br /&gt;
'''Mon Dec  4 11:54:49 EST 2023:''' The nia-datamover1 node will the offline this Monday afternoon for the Globus GCSv5 upgrade. Endpoint computecanada#niagara-GCSv4 will still be available via nia-datamover2&lt;br /&gt;
&lt;br /&gt;
'''Tue Nov 28 16:29:14 EST 2023:''' The computecanada#hpss Globus endpoint is now running GCSv5. We'll find a window of opportunity next week to upgrade computecanada#niagara to GCSv5 as well.&lt;br /&gt;
&lt;br /&gt;
'''Tue Nov 28 14:20:30 EST 2023:''' The computecanada#hpss Globus endpoint will be offline for the next few hours for the GCSv5 upgrade.&lt;br /&gt;
&lt;br /&gt;
'''Fri Nov 10, 2023, 18:00 PM EDT:''' The HPSS upgrade is finished. We didn't have time to update Globus to GCSv5, so we'll find a window of opportunity to do this next week. &lt;br /&gt;
&lt;br /&gt;
Please be advised that starting this &amp;lt;B&amp;gt;Friday morning, Nov/10, we'll be upgrading the HPSS system from version 8.3 to 9.3 and the HPSS Globus server from GCSv4 to GCSv5.&amp;lt;/B&amp;gt; Everything going well we expect to be back online by the end of the day.  &lt;br /&gt;
&lt;br /&gt;
'''Fri Nov 3, 2023, 12:20 PM EDT:''' The &amp;quot;Niagara at Scale&amp;quot; event has finished. Niagara is available again for all users.&lt;br /&gt;
&lt;br /&gt;
'''Tue Oct 31, 2023, 12 PM EDT:''' The &amp;quot;Niagara at Scale&amp;quot; event has started.&lt;br /&gt;
&lt;br /&gt;
'''Tue Oct 31, 2023, 12:PM EDT - Fri Nov 3, 2023, 12:00 PM EDT:''' Three-day reservation for the &amp;quot;Niagara at Scale&amp;quot; event. Only &amp;quot;Niagara at Scale&amp;quot; projects will run on the compute nodes. Users are encouraged to submit small and short jobs that could run before this event.  Throughout the event, users can still login, access their data, and submit jobs, but these jobs will not run until after the event. Note that the debugjob queue will remain available to everyone as well.&lt;br /&gt;
&lt;br /&gt;
''' Thu Oct 27 11:16 AM EDT:''' SSH keys are gradually being restored, estimated to complete by 1:15 PM.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 27, 2023, 8:00 EDT:''' SSH key login authentication with CCDB keys is currently not working, on many Alliance systems.  It appears this started last night. Issue is being investigated.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 26, 2023, 12:35 EDT:''' Mist login node is accessible again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 26, 2023, 12:05 EDT:''' Mist login node is under maintenance and temporarily inaccessible to users.&lt;br /&gt;
&lt;br /&gt;
'''Wed Oct 25 7:54 PM EDT:''' slurm-*.out now outputs job info for last array job.&lt;br /&gt;
&lt;br /&gt;
'''Tue Oct 24 12:00 AM EDT:''' network appears to be up&lt;br /&gt;
&lt;br /&gt;
'''Tue Oct 24 11:32 AM EDT:''' campus network issues&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 05, 2023, 12:05 PM EDT:''' Niagara scheduler is back online.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 05, 2023, 11:50 AM EDT:''' Niagara scheduler is temporarily under maintenance for security updates. &lt;br /&gt;
&lt;br /&gt;
''' Thu Sep 28, 2023 11:00 am''': Niagara scheduler is back online.&lt;br /&gt;
&lt;br /&gt;
''' Thu Sep 28, 2023 10:50 am''': Niagara scheduler is temporarily under maintenance for security updates.&lt;br /&gt;
&lt;br /&gt;
''' Wed Sep 27, 2023 11:35 am''': Mist login node is accessible again.&lt;br /&gt;
&lt;br /&gt;
''' Wed Sep 27, 2023 11:00 am''': Mist login node is under maintenance and temporarily inaccessible to users.&lt;br /&gt;
&lt;br /&gt;
''' Wed Sep 6, 2023 11:30 am''': Mist login node is accessible again.&lt;br /&gt;
&lt;br /&gt;
''' Wed Sep 6, 2023 11:00 am''': Mist login node is under maintenance and temporarily inaccessible to users.&lt;br /&gt;
&lt;br /&gt;
''' Fri Aug 25, 2023 0:19 am''': A power glitch brought some compute nodes down; users should resubmit any affected jobs. The Jupyterhub had to be restarted for the same reason.&lt;br /&gt;
&lt;br /&gt;
''' Mon Aug 14, 2023 12:10 pm''': Network problems with Teach cluster are now resolved and it is again available for users.&lt;br /&gt;
&lt;br /&gt;
''' Mon Aug 14, 2023 11:40 am''': Network problems with Teach cluster. We are investigating.&lt;br /&gt;
&lt;br /&gt;
''' Thu Aug 3, 2023 11:10 am''': Mist login node is accessible again.&lt;br /&gt;
&lt;br /&gt;
''' Thu Aug 3, 2023 10:40 am''': Mist login node is under maintenance and temporarily inaccessible to users.&lt;br /&gt;
&lt;br /&gt;
''' Tue Aug 1, 2023 2:43 pm''': To recover from the power glitch, all servers on the SciNet jupyterhub have been stopped. Please restart you server if you need to.&lt;br /&gt;
&lt;br /&gt;
''' Tue Aug 1, 2023 11:46 am''': There was a power glitch at 11:46 Aug 1, 2023, causing a significant number of job losses. Please resubmit your jobs.&lt;br /&gt;
&lt;br /&gt;
'''Summer Maintenance Shutdown Finished''' -- Slurm upgraded to version 23.02.3.&lt;br /&gt;
Change to be aware: SLURM_NTASKS is only set if --ntasks option is set.&lt;br /&gt;
Details at: https://bugs.schedmd.com/show_bug.cgi?id=17108&lt;br /&gt;
&lt;br /&gt;
'''July 17 and 18, 2023''':  Announcement: Summer Maintenance Shutdown&amp;lt;/span&amp;gt;&amp;lt;br&amp;gt; &lt;br /&gt;
&lt;br /&gt;
'''July 17th, 2023''' This maintenance involves a full data centre shutdown will start at 7:00 a.m. ET on Monday July 17th, 2023. None of the SciNet systems (Niagara, Mist, Rouge, Teach, the file systems, as well as hosted equipment) will be accessible.&lt;br /&gt;
&lt;br /&gt;
'''July 18th, 2023''' The shutdown will last until Tuesday July 18th, 2023. Systems are expected to be fully available in the evening of that day.&lt;br /&gt;
&lt;br /&gt;
The scheduler will hold jobs that cannot finish before the start of the shutdown. Users are encouraged to submit small and short jobs that can take advantage of this, as the scheduler may be able to fit these jobs in before the maintenance on otherwise idle nodes.&lt;br /&gt;
&lt;br /&gt;
'''Wed Jun 21 16:03:45 EDT 2023:''' Niagara's scheduler maintenance is finished.&lt;br /&gt;
&lt;br /&gt;
'''Wed Jun 21 15:42:00 EDT 2023:''' Niagara's scheduler is rebooting in 10 minutes for a short maintenance down time.&lt;br /&gt;
&lt;br /&gt;
'''Wed Jun 21, 2023, 11:25 AM EDT:''' Maintenance is finished and Teach cluster is accessible again.&lt;br /&gt;
&lt;br /&gt;
'''Tue Jun 20, 2023, 9:55 AM EDT:''' Teach cluster is powered off for maintenance.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;span style='color:red'&amp;gt;'''Tue June 20, 2023:'''  Announcement:&amp;lt;/span&amp;gt;&amp;lt;br&amp;gt; The Teach cluster at SciNet will undergo a maintenance shutdown starting on Tuesday June 20, 2023.  It will likely take a few days before it will be available again.  Check here for updates.&lt;br /&gt;
&lt;br /&gt;
'''Mon Jun 5, 2023, 2:35 PM EDT:''' All systems are operational again.&lt;br /&gt;
&lt;br /&gt;
'''Mon Jun 5, 2023, 11:55 AM EDT:''' There were issues with the cooling system.  The login nodes and file systems are now accessible again, but compute nodes are still off.&lt;br /&gt;
&lt;br /&gt;
'''Mon Jun 5, 2023, 6:55 AM EDT:''' Issues at the data center, we are investigating.&lt;br /&gt;
&lt;br /&gt;
'''Sat May 27, 2023, 21:00AM EDT:''' We have been able to mitigate the UPS issue for now, until new parts arrive sometime during the week. System will be accessible soon&lt;br /&gt;
&lt;br /&gt;
'''Sat May 27, 2023, 16:00AM EDT:''' We identified an UPS/Power related issue on the datacenter, that is adversely affecting several components, in particular all file systems. Out of an abundance of caution we are shutting down the cluster, until the UPS situation is resolved. Ongoing jobs will be canceled.&lt;br /&gt;
&lt;br /&gt;
'''Sat May 27, 2023, 11:18AM EDT:''' Filesystem issues, investigating.&lt;br /&gt;
&lt;br /&gt;
'''Wed May 24, 2023, 11:40AM EDT:''' Mist login node is accessible again.&lt;br /&gt;
&lt;br /&gt;
'''Wed May 24, 2023, 11:10 AM EDT:''' Mist login node is under maintenance and temporarily inaccessible to users.&lt;br /&gt;
&lt;br /&gt;
'''Mon May 15, 2023, 10:08 AM EDT''' rebooting Mist-login node again &lt;br /&gt;
&lt;br /&gt;
'''Mon May 15, 2023, 09:15 AM EDT''' rebooting Mist-login node&lt;br /&gt;
&lt;br /&gt;
'''Mon May 01, 2023, 04:00 PM EDT''' done rebooting nia-login nodes&lt;br /&gt;
&lt;br /&gt;
'''Mon May 01, 2023, 12:00 PM EDT''' rebooting all nia-login nodes one at a time &lt;br /&gt;
&lt;br /&gt;
'''Mon May 01, 2023, 11:00 AM EDT''' nia-login07 is going to be rebooted.&lt;br /&gt;
&lt;br /&gt;
'''Thu Apr 20, 2023, 12:05 PM EDT:''' Mist login node is accessible again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Apr 20, 2023, 11:30 AM EDT:''' Mist login node is under maintenance and temporarily inaccessible to users.&lt;br /&gt;
&lt;br /&gt;
'''Thu Apr 20, 2023, 8:27 AM EDT:''' Intermittent file system issues. We are investigating.  For now (10:45 AM), the file systems appear operational.&lt;br /&gt;
&lt;br /&gt;
'''Fri 14 Apr 2023 10:25 AM EDT:''' Switch problem resolved.&lt;br /&gt;
&lt;br /&gt;
'''Fri 14 Apr 2023 10:10 AM EDT:''' A switch problem is affecting access to certain equipment at the SciNet data center, including the Teach cluster.  Niagara and Mist are accessible.&lt;br /&gt;
&lt;br /&gt;
'''Fri 14 Apr 2023 09:55 AM EDT:''' SciNet Jupyter Hub maintenance is finished and it is again available for users.&lt;br /&gt;
&lt;br /&gt;
'''Fri 14 Apr 2023:''' SciNet Jupyter Hub will be restarted for system updates this morning.  Keep in mind to save your notebooks!&lt;br /&gt;
&lt;br /&gt;
'''Thu 06 Apr 2023 03:40 PM EDT:''' Rouge cluster is accessible again.&lt;br /&gt;
&lt;br /&gt;
'''Thu 06 Apr 2023 01:00 PM EDT:''' Rouge cluster is temporarily inaccessible to users due to the electrical work.&lt;br /&gt;
&lt;br /&gt;
'''Sun 02 Apr 2023 03:37 AM EDT:''' IO/read errors on the file system seem to have been fixed. Please resubmit your jobs, and report any further problems to support. Burst Buffer will remain offline for now.&lt;br /&gt;
&lt;br /&gt;
'''Sun 02 Apr 2023 00:18 AM EDT:''' File System is back up, but there seems to be some IO/read errors. All running jobs have been killed. Please hold off on submitting jobs until further notice.&lt;br /&gt;
&lt;br /&gt;
'''Sat 01 Apr 2023 10:17 PM EDT:''' We are having issues with the File System. Currently investigating the cause.&lt;br /&gt;
&lt;br /&gt;
'''Fri 31 Mar 2023 11:00 PM EDT:''' Burst Buffer may be the culprit. We are investigating but may have to take Burst Buffer offline. &lt;br /&gt;
&lt;br /&gt;
'''Fri 31 Mar 2023 01:30 PM EDT:''' File system issues causing trouble for some jobs on Niagara and Mist&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Tue 28 Mar 2023 11:05 AM EDT: Mist login node is accessible again.&lt;br /&gt;
&lt;br /&gt;
Tue 28 Mar 2023 10:35 AM EDT: Mist login node is under maintenance and temporarily inaccessible to users.&lt;br /&gt;
&lt;br /&gt;
Fri 17 Mar 2023 14:50 PM EDT: All systems online.&lt;br /&gt;
&lt;br /&gt;
Fri 17 Mar 2023 11:00 AM EDT: Problem identified and repaired. Starting to bring up systems, but not available to users yet.&lt;br /&gt;
&lt;br /&gt;
Fri 17 Mar 2023 09:15:39 EDT: Staff on site and ticket opened with cooling contractor, cause of failure unclear &lt;br /&gt;
&lt;br /&gt;
Fri 17 Mar 2023 01:47:43 EDT: Cooling system malfunction, datacentre is shut down. &lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Tue Feb 28, 16:40 EST:&amp;lt;/b&amp;gt; All systems are back online.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Tue Feb 28, 15:30 EST:&amp;lt;/b&amp;gt; Maintenance is complete. Bringing up systems.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Tue Feb 28, 7:10 AM EST:&amp;lt;/b&amp;gt; Maintenance shutdown resuming.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Mon Feb 27, 3:55 PM EST:&amp;lt;/b&amp;gt; Maintenance paused as parts were delayed. The maintenance will resume tomorrow (Tue Feb 28) at 7AM EST for about 5 hours.  In the meantime, the login nodes of the systems will be brought online.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Mon Feb 27, 7:20 AM EST:&amp;lt;/b&amp;gt; Maintenance shutdown started.&lt;br /&gt;
 &lt;br /&gt;
&amp;lt;span style=&amp;quot;color:red&amp;quot;&amp;gt;&amp;lt;b&amp;gt; February 27 and 28, 2023: SciNet Data Centre Maintenance:&amp;lt;/b&amp;gt;&amp;lt;/span&amp;gt;&amp;lt;br/&amp;gt;&lt;br /&gt;
This annual winter maintenance involves a full data centre shutdown&lt;br /&gt;
starting at 7:00 a.m. EST on Monday, February 27. None of the SciNet&lt;br /&gt;
systems (Niagara, Mist, Rouge, Teach, the file systems, as well as&lt;br /&gt;
hosted equipment) will be accessible.&lt;br /&gt;
&lt;br /&gt;
On the second day of the maintenance, Niagara, Mist, and their file&lt;br /&gt;
systems are expected to become partially available for users.  All&lt;br /&gt;
systems should be fully available in the evening of the 28th.&lt;br /&gt;
&lt;br /&gt;
The scheduler will hold jobs that cannot finish before the start of&lt;br /&gt;
the shutdown. Users are encouraged to submit small and short jobs&lt;br /&gt;
that can take advantage of this, as the scheduler may be able to fit&lt;br /&gt;
these jobs in before the maintenance on otherwise idle nodes.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Feb 17, 2023, 11:15 PM EST:&amp;lt;/b&amp;gt; File system issues on Teach fixed and Teach is accessible again. Note that the file system of Teach is not very good at handling many remote vscode connections.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Feb 17, 2023, 11:02 PM EST:&amp;lt;/b&amp;gt; File system issues on Teach.  We are working on a fix.  &lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Sun Feb 12, 2023, 3:05 PM EST&amp;lt;/b&amp;gt; All systems are back online.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Sun Feb 12, 2023, 2:10 PM EST&amp;lt;/b&amp;gt; Powers restored, clusters are being started.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Sat Feb 11, 2023, 2:35 PM EST&amp;lt;/b&amp;gt; Powers interruption started. All compute nodes will be down, likely until Sunday &lt;br /&gt;
afternoon.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Sat Feb 11, 2023, 1:20 PM EST&amp;lt;/b&amp;gt; There is to be an emergency power repair on the adjacent street. The datacentre will be &lt;br /&gt;
switching over to generator. All compute nodes will be down.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Fri Feb 10, 2023, 10:55 AM EST&amp;lt;/b&amp;gt; All systems are back online.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Fri Feb 10, 2023, 10:00 AM EST&amp;lt;/b&amp;gt; Cooling issue resolved, cluster is being started.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Wed Jan 25, 2023, 02:15 PM EST&amp;lt;/b&amp;gt; Mist login node is accessible again.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Wed Jan 25, 2023, 10:30 AM EST&amp;lt;/b&amp;gt; Mist login node is under maintenance and temporarily inaccessible to users.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Mon Jan 23, 2023, around 7-8 AM EST&amp;lt;/b&amp;gt; Intermediate file system issuse may have killed your job. Users are advised to resubmit.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Sat Jan 21, 2023, 00:50 EST&amp;lt;/b&amp;gt; Niagara, Mist, Rouge and the filesystems are up&lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Fri Jan 20, 2023, 11:19 PM: EST&amp;lt;/b&amp;gt; Systems are coming up. We have determined that there was a general power glitch in the area of our Datacentre. The power has been fully restored&lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Fri Jan 20, 2023, 10:34 PM: EST&amp;lt;/b&amp;gt; Cooling is back. Systems are slowly coming up  &lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Fri Jan 20, 2023, 8:20 PM: EST&amp;lt;/b&amp;gt; A cooling failure at the data center, possibly due to a power glitch. We are investigating.  &lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;Thu Jan 12, 2023, 9:30 AM EST&amp;lt;/b&amp;gt; File system is experiencing issues. Issues have stabilized, but jobs running around this time may have been affected.&lt;br /&gt;
&lt;br /&gt;
'''Wed Dec 21, 2022, 12:00 PM: ''' Please note that SciNet is on vacation, together with the University of Toronto. Full service will resume on Jan 2, 2023. We will endeavour to keep systems running, and answer tickets, on a best-effort basis.  Happy Holidays!!!&lt;br /&gt;
&lt;br /&gt;
'''Fri Dec 16, 2022, 2:19 PM: ''' City power glitch caused all compute nodes to reboot. Please resubmit your jobs.&lt;br /&gt;
&lt;br /&gt;
'''Mon Dec 12, 2022, 9:30 AM - 11:30:''' File system issues caused login issues and may have affected running jobs.  System back to normal now, but users may want to check any jobs they had running. &lt;br /&gt;
&lt;br /&gt;
'''Wed Dec 7, 2022, 11:40 AM EST:''' Systems are being brought back online.&lt;br /&gt;
&lt;br /&gt;
'''Wed Dec 7, 2022, 09:00 AM EST:''' Maintenance is underway.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;span style='color:red'&amp;gt;&amp;lt;b&amp;gt;Announcement:&amp;lt;/b&amp;gt;&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
On '''Wednesday December 7th, 2022''', the file systems of the SciNet's systems, Niagara, Mist, HPSS, Teach cluster, will undergo maintenance from 9:00 am EST.  During the maintenance, there will be no access to any of these systems, as it requires all file system operations to have stopped.  The maintenance should take about 1 hour, and all systems are expected to become available again later that morning.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
'''Wed Nov 30, 2022, 14:45 PM EST:''' Mist login node is accessible again.&lt;br /&gt;
&lt;br /&gt;
'''Wed Nov 30, 2022, 14:15 PM EST:''' Mist login node is under maintenance and temporarily inaccessible to users. &lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 20, 2022, 18:00 PM EDT:''' Systems are back online &lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 20, 2022, 09:40 AM EDT:''' About half of Niagara compute nodes are up. Note that only jobs that can finish by 5:00 PM will run.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 20, 2022, 07:50 AM EDT:''' Jupyter Hub is available again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 20, 2022, 07:35 AM EDT:''' Jupyter Hub is being updated and temporarily inaccessible to users.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 20, 2022, 07:30 AM EDT:''' Maintenance is underway.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;span style='color:red'&amp;gt;&amp;lt;b&amp;gt;Announcement:&amp;lt;/b&amp;gt;&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
On '''Thursday October 20th, 2022''', the SciNet datacentre (which hosts Niagara and Mist) will undergo transformer maintenance from 7:30 am EDT to 5:00 pm EDT.  At both the start and end of this maintenance window, all systems will need to be briefly shutdown and will not be accessible.  Apart from that, during this window, login nodes will be accessible and part of Niagara will be available to run jobs. The Mist and Rouge clusters will be off for the entirety of this maintenance. &lt;br /&gt;
&lt;br /&gt;
Users are encouraged to submit Niagara jobs of about 1 to 2 hours in the days before the maintenance, as these could be run within the&lt;br /&gt;
window of 8 AM and 5 PM EDT.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
'''Wed Oct 5, 2022, 12:10 PM EDT:''' A grid power glitch caused all compute nodes to reboot. Please resubmit your jobs.&lt;br /&gt;
&lt;br /&gt;
'''Mon Oct 3, 2022, 11:20 PM EDT:'''  Niagara login nodes are accessible from outside again.&lt;br /&gt;
&lt;br /&gt;
'''Mon Oct 3, 2022, 9:20 PM EDT:'''  Niagara login nodes are inaccessible from outside of the datacentre at the moment. As a work-around, ssh into mist.scinet.utoronto.ca and then ssh into e.g. nia-login01.&lt;br /&gt;
&lt;br /&gt;
'''Wed Sep 28, 2022, 1:15 PM EDT:''' The JupyterHub maintenance is finished and it is now accessible to users.&lt;br /&gt;
&lt;br /&gt;
'''Wed Sep 28, 2022, 1:00 PM EDT:''' The JupyterHub is to be rebooted for system upgrades. Running processes and notebooks will be closed. The service is expected to be back around 1:30 PM EDT.&lt;br /&gt;
 &lt;br /&gt;
'''Tue Sep 27, 2022, 11:50 AM EDT:''' Mist login node is accessible again.&lt;br /&gt;
&lt;br /&gt;
'''Tue Sep 27, 2022, 11:25 AM EDT:''' Mist login node is under maintenance and temporarily inaccessible to users.&lt;br /&gt;
&lt;br /&gt;
'''Mon Sep 26, 2022, 11:35 AM EDT:''' Rouge and Teach login nodes are accessible again.&lt;br /&gt;
&lt;br /&gt;
'''Mon Sep 26, 2022, 11:05 AM EDT:''' Rouge and Teach login nodes are under maintenance and temporarily inaccessible to users.&lt;br /&gt;
&lt;br /&gt;
'''Fri Sep 22, 2022, 0:46 AM EDT:''' The CCEnv software stack is back to normal.&lt;br /&gt;
&lt;br /&gt;
'''Thu Sep 22, 2022, 8:15 PM EDT:''' The CCEnv software stack is inaccessible due to an issue with CVMFS.&lt;br /&gt;
 &lt;br /&gt;
'''Tue Sep 20, 2022, 16:00 AM EDT:''' Rouge login node is accessible again.&lt;br /&gt;
&lt;br /&gt;
'''Tue Sep 20, 2022, 10:20 AM EDT:''' Rouge login node is under maintenance and temporarily inaccessible to users (hardware upgrade).&lt;br /&gt;
&lt;br /&gt;
'''Tue Sep 20, 2022, 9:41 AM EDT:''' Rouge login node is back up.&lt;br /&gt;
&lt;br /&gt;
'''Tue Sep 20, 2022, 8:25 AM EDT:''' Rouge login node down, we are investigating.&lt;br /&gt;
&lt;br /&gt;
'''Fri Sept 16, 2022, 9:30 AM EDT:''' Login nodes are accessible again.&lt;br /&gt;
&lt;br /&gt;
'''Fri Sept 16, 2022, 9:00 AM EDT:''' Login nodes are not accessible.  We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''Tue Sep 13, 2022, 11:00 AM EDT:''' Mist login node is available again.&lt;br /&gt;
&lt;br /&gt;
'''Tue Sep 13, 2022, 10:00 AM EDT:''' Mist login node is under maintenance and temporarily inaccessible to users.&lt;br /&gt;
&lt;br /&gt;
'''Fri Sep 2, 2022, 11:25 AM EDT:''' Rouge login node is back up.&lt;br /&gt;
&lt;br /&gt;
'''Fri Sep 2, 2022, 10:25 AM EDT:''' Issues with the Rouge login node; we are investigating.&lt;br /&gt;
&lt;br /&gt;
'''Tue Aug 23, 2022, 1:15 PM EDT:''' Jupyter Hub is available again.&lt;br /&gt;
&lt;br /&gt;
'''Tue Aug 23, 2022, 1:00 PM EDT:''' Jupyter Hub is being updated and temporarily inaccessible to users.&lt;br /&gt;
&lt;br /&gt;
'''Fri Aug 12, 2022, 6:30 PM EDT:''' File system issues are resolved.&lt;br /&gt;
&lt;br /&gt;
'''Fri Aug 12, 2022, 5:06 PM EDT:''' File system issues. We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''Thu Aug 11, 2022, 9:20 AM EDT:''' The login node issues have been resolved.&lt;br /&gt;
&lt;br /&gt;
'''Thu Aug 11, 2022, 7:50 AM EDT:''' We are having problems accessing the Niagara login nodes.  Until fixed, please login to Mist and then ssh to a Niagara login node to access Niagara (&amp;quot;ssh nia-login02&amp;quot;, for example).&lt;br /&gt;
&lt;br /&gt;
'''Fri July 15, 2022, 10:50 AM EDT:''' Jupyter Hub is available again.&lt;br /&gt;
&lt;br /&gt;
'''Fri July 15, 2022, 10:30 AM EDT:''' Jupyter Hub is being updated and temporarily inaccessible to users.&lt;br /&gt;
&lt;br /&gt;
'''Wed June 16, 2022, 3:45 PM EDT:''' File system is stable now. We're gradually opening the systems up.&lt;br /&gt;
&lt;br /&gt;
'''Wed June 16, 2022, 10:15 AM EDT:''' Emergency maintenance shutdown of filesystem. Running jobs will be affected.&lt;br /&gt;
&lt;br /&gt;
'''Wed June 15, 2022, 7:35 PM EDT:''' Maintenance shutdown finished. Most systems are available again.&lt;br /&gt;
&lt;br /&gt;
'''Wed June 15, 2022, 7:00 AM EDT:''' Maintenance shutdown of the SciNet datacentre. There will be no access to any of the SciNet systems during this time. We expect to be able to bring the systems back online in the evening of June 15th.&lt;br /&gt;
&lt;br /&gt;
'''Mon June 13, 2022, 7:00 AM EDT - Wed June 15, 2022, 7:00 AM EDT:''' Two-day reservation for the &amp;quot;Niagara at Scale&amp;quot; event. Only &amp;quot;Niagara at Scale&amp;quot; projects will run on the compute nodes (as well as SOSCIP projects, on a subset of nodes). Users are encouraged to submit small and short jobs that could run before this event.  Throughout the event, users can still login, access their data, and submit jobs, but these jobs will not run until after the subsequent maintenance (see below). Note that the debugjob queue will remain available to everyone as well.&lt;br /&gt;
&lt;br /&gt;
'''Mon May 30th, 2022, 12:42:00 EDT:''' Mist login node is available again.&lt;br /&gt;
&lt;br /&gt;
'''Mon May 30th, 2022, 10:22:00 EDT:''' Mist login node is being upgraded and temporarily inaccessible to users.&lt;br /&gt;
&lt;br /&gt;
'''Wed May 25th, 2022, 13:30:00 EDT:''' Niagara operating at 100% again.&lt;br /&gt;
&lt;br /&gt;
'''Tue May 24th, 2022, 21:30:00 EDT:''' Jupyter Hub up.  Part of Niagara can run compute jobs.&lt;br /&gt;
&lt;br /&gt;
'''Tue May 24th, 2022, 19:00:00 EDT:''' Systems are up. Users can login, BUT cannot submit jobs yet.&lt;br /&gt;
&lt;br /&gt;
'''Tue May 24th, 2022, 10:00:00 EDT:''' We are still performing system checks.&lt;br /&gt;
&lt;br /&gt;
'''Mon May 23rd, 2022, 16:44:30 EDT:''' Systems still down. Filesystems are working, but there are quite a number of drive failures - no data loss - so out of an abundance of caution we are keeping the systems down at least until tomorrow.  The long weekend has also been disruptive for service response, and we prefer to err on the safe side.&lt;br /&gt;
&lt;br /&gt;
'''Mon May 23rd, 2022, 08:12:14 EDT:''' Systems still down. Filesystems being checked to ensure no heat damage.&lt;br /&gt;
&lt;br /&gt;
'''Sun May 22nd, 2022, 10.16 am EDT:''' Electrician dispatched to replace blown fuses.&lt;br /&gt;
&lt;br /&gt;
'''Sun May 22nd, 2022, 2:54 am EDT:''' Automatic shutdown down due to power/cooling.&lt;br /&gt;
&lt;br /&gt;
'''Fri May 6th, 2022, 11:35 am EDT:''' HPSS scheduler upgrade also finished.&lt;br /&gt;
&lt;br /&gt;
'''Thu May 5th, 2022, 7:45 pm EDT:''' Upgrade of the scheduler has finished, with the exception of HPSS.&lt;br /&gt;
&lt;br /&gt;
'''Thu May 5th, 2022, 7:00 am - 3:00 pm EDT (approx):''' Starting from 7:00 am EDT, an upgrade of the scheduler of the Niagara, Mist, and Rouge clusters will be applied.  This requires the scheduler to be down for about 5-6 hours, and all compute and login nodes to be rebooted.&lt;br /&gt;
Jobs cannot be submitted during this maintenance, but jobs submitted beforehand will remain in the queue.  For most of the time, the login nodes of the clusters will be available so that users may access their files on the home, scratch, and project file systems.&lt;br /&gt;
&lt;br /&gt;
'''Monday May 2nd, 2022, 9:30 - 11:00 am EDT:''' the Niagara login nodes, the jupyter hub, and nia-datamover2 will get rebooted for updates.  In the process, any login sessions will get disconnected, and servers on the jupyterhub will stop. Jobs in the Niagara queue will not be affected.&lt;br /&gt;
&lt;br /&gt;
'''Tue Apr 26, 11:20 AM EDT:''' A Rolling update of the Mist cluster is taking a bit longer than expected, affecting logins to Mist. &lt;br /&gt;
 &lt;br /&gt;
'''Announcement:''' On Thursday April 14th, 2022, the connectivity to the SciNet datacentre will be disrupted at 11:00 AM EDT  for a few minutes, in order to deploy a new network core switch.  Any SSH connections or data transfers to SciNet systems (Niagara, Mist, etc.) may be terminated at that time.&lt;br /&gt;
&lt;br /&gt;
'''Thu March 24, 6:54 AM EST:''' HPSS is back online&lt;br /&gt;
&lt;br /&gt;
'''Thu March 24, 8:15 AM EST:''' HPSS has a hardware problem&lt;br /&gt;
&lt;br /&gt;
'''Wed March 2, 4:50 PM EST:''' The CCEnv software stack is available again on Niagara.&lt;br /&gt;
&lt;br /&gt;
'''Wed March 2, 7:50 AM EST:''' The CCEnv software stack on Niagara has issues; we are investigating.&lt;br /&gt;
 &lt;br /&gt;
'''Sat Feb 12 2022, 12:59 EST:''' Jupyterhub is back up, but may have hardware issue.&lt;br /&gt;
&lt;br /&gt;
'''Sat Feb 12 2022, 10:36 EST:''' Issue with the Jupyterhub, since last night.  We're investigating.&lt;br /&gt;
&lt;br /&gt;
'''Tue Feb 1 2022 19:20 EST:''' Maintenance finished successfully. Systems are up. &lt;br /&gt;
&lt;br /&gt;
'''Tue Feb 1 2022 13:00 EST:''' Maintenance downtime started.&lt;br /&gt;
&lt;br /&gt;
'''Mon Jan 31 2022 13:15:00 EST:''' The SciNet datacentre's cooling system needs an '''emergency repair''' as soon as possible.  During this repair, all systems hosted at SciNet (Niagara, Mist, Rouge, HPSS, and Teach) will need to be switched off and will be unavailable to users. Repairs will start '''Tuesday February 1st, at 1:00 pm EST''', and could take until the end of the next day.  Please check here for updates.&lt;br /&gt;
&lt;br /&gt;
'''Sat Jan 29 2020 16:45:38 EST:''' Fibre repaired.&lt;br /&gt;
&lt;br /&gt;
'''Sat 29 Jan 2022 11:22:27 EST:''' Fibre repair is underway.  Expect to have connectivity restored later today.&lt;br /&gt;
&lt;br /&gt;
'''Fri 28 Jan 2022 07:35:01 EST:''' The fibre optics cable that connects the SciNet datacentre was severed by uncoordinated digging at York University.  We expect repairs to happen as soon as possible.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 27 12:46 EST PM 2022:''' Network issues to and from the datacentre. We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''Sun Jan 23 11:05 EST AM 2022:''' Filesystem issues appear to have resolved.&lt;br /&gt;
&lt;br /&gt;
'''Sun Jan 23 10:30 EST AM 2022:''' Filesystem issues -- investigating.&lt;br /&gt;
&lt;br /&gt;
'''Sat Jan 8 11:42 EST AM 2022:''' The emergency maintenance is complete. Systems are up and available.&lt;br /&gt;
&lt;br /&gt;
'''Fri Jan 7 14:34 EST PM 2022:''' The SciNet shutdown is in progress. Systems are expected back on Saturday, Jan 8.&lt;br /&gt;
&lt;br /&gt;
'''&amp;lt;span style=&amp;quot;color:red&amp;quot;&amp;gt;Emergency shutdown Friday January 7, 2022&amp;lt;/span&amp;gt;''': An emergency shutdown of all SciNet to replace a crucial file system component is planned to take place on Friday January 7, 2022, starting at 8am EST, and will require at least 12 hours of downtime.  Updates will be posted during the day.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 6 08:20 EST AM 2022''' The SciNet filesystem is having issues.  We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''Fri Dec 24 13:31 EST PM 2021''' Please note the following scheduled network maintenance, which will result in loss of connectivity to the SciNet datacentre:  Start time&lt;br /&gt;
Dec 29, 00:30 EST  Estimated duration  4 hours and 30 minutes. &lt;br /&gt;
&lt;br /&gt;
'''Mon Dec 20 4:29 EST PM 2021''' Filesystem is back to normal. &lt;br /&gt;
&lt;br /&gt;
'''Mon Dec 20 2:53 EST PM 2021''' Filesystem problem - We are investigating. &lt;br /&gt;
&lt;br /&gt;
'''Wed Sep 23 12:30 EDT 2021 ''' Cooling restored.  Systems should be available later this afternoon.  &lt;br /&gt;
&lt;br /&gt;
'''Wed Sep 23 9:30 EDT 2021 ''' Technicians on site working on cooling system. &lt;br /&gt;
&lt;br /&gt;
'''Wed Sep 23 3:30 EDT 2021 ''' Cooling system issues still unresolved. &lt;br /&gt;
&lt;br /&gt;
'''Wed Sep 22 23:27:48 EDT 2021 ''' Shutdown of the datacenter due to a problem with the cooling system.&lt;br /&gt;
&lt;br /&gt;
'''Wed Sep 22 09:30 EDT 2021 ''': File system issues, resolved.&lt;br /&gt;
&lt;br /&gt;
'''Wed Sep 22 07:30 EDT 2021 ''': File system issues, investigating.&lt;br /&gt;
&lt;br /&gt;
'''Sun Sep 19 10:00 EDT 2021''': Power glitch interrupted all compute jobs; please resubmit any jobs you had running.&lt;br /&gt;
&lt;br /&gt;
'''Wed Sep 15 17:35 EDT 2021''': filesystem issues resolved&lt;br /&gt;
&lt;br /&gt;
'''Wed Sep 15 16:39 EDT 2021''': filesystem issues&lt;br /&gt;
&lt;br /&gt;
'''Mon Sep 13 13:15:07 EDT 2021''' HPSS is back online.&lt;br /&gt;
&lt;br /&gt;
'''Fri Sep 10 17:57:23 EDT 2021''' HPSS is offline due to unscheduled maintenance.&lt;br /&gt;
&lt;br /&gt;
'''Wed Aug 18 16:13:42 EDT 2021''' The HPSS upgrade is complete.&lt;br /&gt;
&lt;br /&gt;
'''HPSS Downtime August 17th and 18th, 2021 (Tuesday and Wednesday):''' We'll be upgrading the HPSS software to version 8.3, along with all the clients (htar/hsi, vfs and Globus/dsi)&lt;br /&gt;
&lt;br /&gt;
'''July 24, 2021, 6:00 PM EDT:''' There appear to be file system issues, which may affect users' ability to login.  We are investigating.&lt;br /&gt;
&lt;br /&gt;
''' July 23th, 2021, 9:00 AM EDT:''' ''' Security update: ''' Due to a severe vulnerability in the Linux kernel (CVE-2021-33909), our team is currently patching and rebooting all login nodes and compute nodes, as well as the JupyterHub.  There should be no affect on running jobs, however sessions on login and datamover nodes will be disrupted. &lt;br /&gt;
&lt;br /&gt;
''' July 20th, 2021, 7:00 PM EDT:''' ''' SLURM configuration''' - Changed the default behaviour to kill a job step if any task exits with a non-zero exit code. If your code is able to handle failures gracefully, please add srun's option --no-kill to recover the previous default behaviour.&lt;br /&gt;
&lt;br /&gt;
''' July 20th, 2021, 7:00 PM EDT:''' Maintenance finished, systems are back online.   &lt;br /&gt;
&lt;br /&gt;
'''SciNet Downtime July 20th, 2021 (Tuesday):''' There will be a maintenance shutdown of the SciNet data center on Tuesday July 20th, starting at 7 am EDT. There will be no access to any of the SciNet systems (Niagara, Mist, HPSS, Teach cluster, or the file systems) during this time.  We expect to be able to bring the systems back online in the evening of July 20th.  The status of the Niagara cluster can be checked on status.computecanada.ca. For up-to-date and more detailed information on the status of all the SciNet systems, you can always check back here.&lt;br /&gt;
&lt;br /&gt;
'''June 28th, 2021, 4:06 PM:''' Mist OS upgrade is complete.&lt;br /&gt;
&lt;br /&gt;
'''May 27, 2021:''' Datamovers addresses have changed to improve high bandwidth connectivity and cybersecurity. The new addresses are 142.1.174.227 for nia-datamover1.scinet.utoronto.ca, and 142.1.174.228 for nia-datamover2.scinet.utoronto.ca.&lt;br /&gt;
&lt;br /&gt;
If you have jobs that need to connect to a software license server using an ssh tunnel through nia-gw (which actually resolves to datamover1 or datamover2), you may need to ask the system administrators of that license server to allow incoming connections from the new addresses above.&lt;br /&gt;
'''June 29th, 2021, 2:00 PM:''' Thunderstorm-related power fluctuations are causing some Niagara compute nodes and their jobs to crash.  Please resubmit if your jobs seem to have crashed for no apparent reason.&lt;br /&gt;
&lt;br /&gt;
'''June 28th, 2021, 4:06 PM:''' Mist OS upgrade is complete.&lt;br /&gt;
&lt;br /&gt;
'''June 28th, 2021, 9:00 AM:''' Mist is under maintenance. OS upgrading from RHEL 7 to 8.&lt;br /&gt;
&lt;br /&gt;
'''June 11th, 2021, 8:30 AM:''' Maintenance complete. Systems are up.&lt;br /&gt;
&lt;br /&gt;
'''June 9th to 10th, 2021:''' The SciNet datacentre will have a scheduled maintenance shutdown.  Niagara, Mist, Rouge, HPSS, login nodes, the file systems, and hosted systems will all be offline during the shutdown starting at 7AM EDT on Wednesday June 9th. We expect the systems to be back up in the morning of Friday June 11th.  Check here for updates.&lt;br /&gt;
&lt;br /&gt;
'''May 27, 2021:''' Datamovers addresses have changed to improve high bandwidth connectivity and cybersecurity. The new addresses are 142.1.174.227 for nia-datamover1.scinet.utoronto.ca, and 142.1.174.228 for nia-datamover2.scinet.utoronto.ca.&lt;br /&gt;
&lt;br /&gt;
'''May 27th, 20:00.''' All systems are up and running &lt;br /&gt;
&lt;br /&gt;
'''May 27th, 19:30.''' Most systems are up&lt;br /&gt;
&lt;br /&gt;
'''May 27th, 19:00:''' Cooling is back. Powering up systems&lt;br /&gt;
&lt;br /&gt;
'''May 27th, 2021, 11:30am:'''  The cooling tower issue has been identified as a wiring issue and is being repaired.  We don't have an ETA on when cooling will be restored, however we are hopeful it will be by the end of the day.  &lt;br /&gt;
&lt;br /&gt;
'''May 27th, 2021, 12:30am:''' Cooling tower motor is not working properly and may need to be replaced.  Its the primary motor and the cooling system can not run without it, so at least until tomorrow all equipment at the datacenter will remain unavailable.  Updates about expected repair times will be posted when they are known.&lt;br /&gt;
&lt;br /&gt;
'''May 26th, 2021, 9:20pm:''' we are currently experiencing cooling issues at the SciNet data centre.  Updates will be posted as we determine the cause of the problem.&lt;br /&gt;
&lt;br /&gt;
'''From Tue Mar 30 at 12 noon EST to Thu Apr 1 at 12 noon EST,''' there will be a two-day reservation for the &amp;quot;Niagara at Scale&amp;quot; pilot  event.  During these 48 hours, only &amp;quot;Niagara at Scale&amp;quot; projects will run on the compute notes (as well as SOSCIP projects, on a subset of nodes).  All other users can still login, access their data, and submit jobs throughout this event, but the jobs will not run until after the event.  The debugjob queue will remain available to  everyone as well.&lt;br /&gt;
&lt;br /&gt;
The scheduler will not start batch jobs that cannot finish before the start of this event. Users can submit small and short jobs can take advantage of this, as the scheduler may be able to fit these jobs in before the event starts on the otherwise idle nodes.&lt;br /&gt;
&lt;br /&gt;
'''Tue 23 Mar 2021 12:19:07 PM EDT''' - Planned external network maintenance 12pm-1pm Tuesday, March 23rd. &lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 28 17:35:16 EST 2021:''' HPSS services are back online&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 28 12:36:21 EST 2021:''' HPSS services offline&lt;br /&gt;
&lt;br /&gt;
We need a small maintenance window as early as possible still this afternoon to perform a small change in configuration. Ongoing jobs will be allowed to finish, but we are keeping new submissions on hold on the queue.&lt;br /&gt;
&lt;br /&gt;
'''Mon Jan 25 13:16:33 EST 2021:''' HPSS services are back online&lt;br /&gt;
&lt;br /&gt;
'''Sat Jan 23 10:03:33 EST 2021:''' HPSS services offline&lt;br /&gt;
&lt;br /&gt;
We detected some type of hardware failure on our HPSS equipment overnight, so access has been disabled pending further investigation.&lt;br /&gt;
&lt;br /&gt;
'''Fri Jan 22 10:49:29 EST 2021:''' The Globus transition to oauth is finished&lt;br /&gt;
&lt;br /&gt;
Please deactivate any previous sessions to the niagara endpoint (in the last 7 days), and activate/login again. &lt;br /&gt;
&lt;br /&gt;
For more details check https://docs.scinet.utoronto.ca/index.php/Globus#computecandada.23niagara&lt;br /&gt;
&lt;br /&gt;
'''Jan 21, 2021:''' Globus access disruption on Fri, Jan/22/2021 10AM: Please be advised that we will have a maintenance window starting tomorrow at 10AM to roll out the transition of services to oauth based authentication.&lt;br /&gt;
&lt;br /&gt;
'''Jan 15, 2021:'''Globus access update on Mon, Jan/18/2021 and Tue, Jan/19/2021:&lt;br /&gt;
Please be advised we start preparations on Monday to perform update to Globus access on Tuesday. We'll be adopting oauth instead of myproxy from that point on. During this period expect sporadic disruptions of service. On Monday we'll already block access to nia-dm2, so please refrain from starting new login sessions or ssh tunnels via nia-dm2 from this weekend already.&lt;br /&gt;
&lt;br /&gt;
''' December 11,2020, 12:00 AM EST: ''' Cooling issue resolved. Systems back.&lt;br /&gt;
&lt;br /&gt;
''' December 11,2020, 6:00 PM EST: ''' Cooling issue at datacenter. All systems down.&lt;br /&gt;
&lt;br /&gt;
''' December 7, 2020, 7:25 PM EST: '''All systems back; users can log in again.&lt;br /&gt;
&lt;br /&gt;
''' December 7, 2020, 6:46 PM EST: '''User connectivity to data center not yet ready, but queued jobs on Mist and Niagara have been started.&lt;br /&gt;
 &lt;br /&gt;
''' December 7, 2020, 7:00 AM EST: '''Maintenance shutdown in effect. This is a one-day maintenance shutdown.  There will be no access to Niagara, Mist, HPSS or teach, nor to their file systems during this time.  We expect to be able to bring the systems back online this evening.&lt;br /&gt;
&lt;br /&gt;
''' December 2, 2020, 9:10 PM EST: '''Power is back, systems are coming up. Please resubmit any jobs that failed because of this incident.&lt;br /&gt;
&lt;br /&gt;
''' December 2, 2020, 6:00 PM EST: '''Power glitch at the data center, caused about half of the compute nodes to go down.  Power issue not yet resolved.&lt;br /&gt;
&lt;br /&gt;
'''&amp;lt;span style=&amp;quot;color:#dd1111&amp;quot;&amp;gt;Announcing a Maintenance Shutdown on December 7th, 2020&amp;lt;/span&amp;gt;''' &amp;lt;br/&amp;gt;There will be a one-day maintenance shutdown on December 7th 2020, starting at 7 am EST.  There will be no access to Niagara, Mist, HPSS or teach, nor to their file systems during this time.  We expect to be able to bring the systems back online in the evening of the same day.&lt;br /&gt;
&lt;br /&gt;
''' November 6, 2020, 8:00 PM EST: ''' Systems are coming back online.&lt;br /&gt;
&lt;br /&gt;
''' November 6, 2020, 9:49 AM EST: ''' Repairs on the cooling system are underway.  No ETA, but the systems will likely be back some time today.&lt;br /&gt;
&lt;br /&gt;
''' November 6, 2020, 4:27 AM EST: '''Cooling system failure, datacentre is shut down.&lt;br /&gt;
&lt;br /&gt;
''' October 9, 2020, 12:57 PM: ''' A short power glitch caused many of the Niagara compute nodes to lose power; jobs running on them would have failed. Please check your jobs and resubmit.&lt;br /&gt;
&lt;br /&gt;
''' October 8, 2020, 9:50 PM: ''' Jupyterhub service is back up.&lt;br /&gt;
&lt;br /&gt;
''' October 8, 2020, 5:40 PM: ''' Jupyterhub service is down. We are investigating.&lt;br /&gt;
&lt;br /&gt;
''' September 28, 2020, 11:00 AM EST: ''' A short power glitch caused many of the Niagara compute nodes to lose power; jobs running on them would have failed. Please check your jobs and resubmit.&lt;br /&gt;
&lt;br /&gt;
''' September 1, 2020, 2:15 PM EST: ''' A short power glitch caused about half of the Niagara compute nodes to lose power; jobs running on them would have failed. Please check your jobs and resubmit.&lt;br /&gt;
&lt;br /&gt;
''' September 1, 2020, 9:27 AM EST: ''' The Niagara cluster has moved to a new default software stack, NiaEnv/2019b.  If your job scripts used the previous default software stack before (NiaEnv/2018a), please put the command &amp;quot;module load NiaEnv/2018a&amp;quot; before other module commands in those scripts, to ensure they will continue to work, or try the new stack (recommended).&lt;br /&gt;
''' August 24, 2020, 7:37 PM EST: ''' Connectivity is back to normal&lt;br /&gt;
&lt;br /&gt;
''' August 24, 2020, 6:35 PM EST: ''' We have partial connectivity back, but are still investigating.&lt;br /&gt;
&lt;br /&gt;
''' August 24, 2020, 3:15 PM EST: ''' There are issues connecting to the data centre. We're investigating.&lt;br /&gt;
&lt;br /&gt;
''' August 21, 2020, 6:00 PM EST: ''' The pump has been repaired, cooling is restored, systems are up.  &amp;lt;br/&amp;gt;Scratch purging is postponed until the evening of Friday Aug 28th, 2020.&lt;br /&gt;
&lt;br /&gt;
'''August 19, 2020, 4:40 PM EST:''' Update: The current estimate is to have the cooling restored on Friday and we hope to have the systems available for users on Saturday August 22, 2020.&lt;br /&gt;
&lt;br /&gt;
'''August 17, 2020, 4:00 PM EST:''' Unfortunately after taking the pump apart it was determined there was a more serious failure of the main drive shaft, not just the seal. As a new one will need to be sourced or fabricated we're estimating that it will take at least a few more days to get the part and repairs done to restore cooling. Sorry for the inconvenience. &lt;br /&gt;
&lt;br /&gt;
'''August 15, 2020, 1:00 PM EST:''' Due to parts availablity to repair the failed pump and cooling system it is unlikely that systems will be able to be restored until Monday afternoon at the earliest. &lt;br /&gt;
&lt;br /&gt;
'''August 15, 2020, 00:04 AM EST:'''  A primary pump seal in the cooling infrastructure has blown and parts availability will not be able be determined until tomorrow. All systems are shut down as there is no cooling.  If parts are available, systems may be back at the earliest late tomorrow. Check here for updates.  &lt;br /&gt;
&lt;br /&gt;
'''August 14, 2020, 21:04 AM EST:''' Tomorrow's /scratch purge has been postponed.&lt;br /&gt;
&lt;br /&gt;
'''August 14, 2020, 21:00 AM EST:''' Staff at the datacenter. Looks like one of the pumps has a seal that is leaking badly.&lt;br /&gt;
&lt;br /&gt;
'''August 14, 2020, 20:37 AM EST:''' We seem to be undergoing a thermal shutdown at the datacenter.&lt;br /&gt;
&lt;br /&gt;
'''August 14, 2020, 20:20 AM EST:''' Network problems to niagara/mist. We are investigating.&lt;br /&gt;
 &lt;br /&gt;
'''August 13, 2020, 10:40 AM EST:''' Network is fixed, scheduler and other services are back.&lt;br /&gt;
&lt;br /&gt;
'''August 13, 2020, 8:20 AM EST:''' We had an IB switch failure, which is affecting a subset of nodes, including the scheduler nodes.&lt;br /&gt;
&lt;br /&gt;
'''August 10, 2020, 7:30 PM EST:''' Scheduler fully operational again.&lt;br /&gt;
&lt;br /&gt;
'''August 10, 2020, 3:00 PM EST:''' Scheduler partially functional: jobs can be submitted and are running.&lt;br /&gt;
&lt;br /&gt;
'''August 10, 2020, 2:00 PM EST:''' Scheduler is temporarily inoperational.&lt;br /&gt;
&lt;br /&gt;
'''August 7, 2020, 9:15 PM EST:''' Network is fixed, scheduler and other services are coming back.&lt;br /&gt;
&lt;br /&gt;
'''August 7, 2020, 8:20 PM EST:''' Disruption of part of the network in the data centre.  Causes issue with the scheduler, the mist login node, and possibly others. We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''July 30, 2020, 9:00 AM''' Project backup in progress but incomplete: please be aware that after we deployed the new, larger storage appliance for scratch and project two months ago, we started a full backup of project (1.5PB). This backup is taking a while to complete, and there are still a few areas which have not been backed up fully. Please be careful to not delete things from project that you still need, in particular if they are recently added material.&lt;br /&gt;
&lt;br /&gt;
'''July 27, 2020, 5:00 PM:''' Scheduler issues resolved.&lt;br /&gt;
&lt;br /&gt;
'''July 27, 2020, 3:00 PM:''' Scheduler issues. We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''July 13, 4:40 PM:''' Most systems are available again. Only Mist is still being brought up.&lt;br /&gt;
&lt;br /&gt;
'''July 13, 10:00 AM:''' '''SciNet/Niagara Downtime In Progress'''&lt;br /&gt;
&lt;br /&gt;
'''SciNet/Niagara Downtime Announcement, July 13, 2020'''&amp;lt;br/&amp;gt;&lt;br /&gt;
All resources at SciNet will undergo a maintenance shutdown on Monday July 13, 2020, starting at 10:00 am EDT, for file system and scheduler upgrades.  There will be no access to any of the SciNet systems (Niagara, Mist, HPSS, Teach cluster, or the file systems) during this time.&lt;br /&gt;
We expect to be able to bring the systems back around 3 PM (EST) on the same day.&lt;br /&gt;
&lt;br /&gt;
''' June 29, 6:21:00  PM:''' Systems are available again.  &lt;br /&gt;
&lt;br /&gt;
''' June 29, 12:30:00  PM:''' Power Outage caused thermal shutdown.&lt;br /&gt;
&lt;br /&gt;
'''June 20, 2020, 10:24 PM:''' File systems are back up.  Unfortunately, all running jobs would have died and users are asked to resubmit them.&lt;br /&gt;
&lt;br /&gt;
'''June 20, 2020, 9:48 PM:''' An issue with the file systems is causing trouble.  We are investigating the cause.&lt;br /&gt;
&lt;br /&gt;
'''June 15, 2020, 10:30 PM:''' A '''power glitch''' caused some compute nodes to be rebooted: jobs running at the time may have failed; users are asked to resubmit these jobs.&lt;br /&gt;
&lt;br /&gt;
'''June 12, 2020, 6:15 PM:''' Two '''power glitches''' during the night caused some compute nodes to be rebooted: jobs running at the time may have failed; users are asked to resubmit these jobs.&lt;br /&gt;
&lt;br /&gt;
'''June 6, 2020, 6:06 AM:''' A '''power glitch''' caused some compute nodes to be rebooted: jobs running at the time may have failed; users are asked to resubmit these jobs.&lt;br /&gt;
&lt;br /&gt;
'''May 24, 2020, 8:20 AM:''' A '''power glitch''' this morning caused all compute nodes to be rebooted: jobs running at the time may have failed; users are asked to resubmit these jobs.&lt;br /&gt;
&lt;br /&gt;
'''May 7, 2020, 6:05 PM:''' Maintenance shutdown is finished.  Most systems are back in production.&lt;br /&gt;
&lt;br /&gt;
'''May 6, 2020, 7:08 AM:''' Two-day datacentre maintenance shutdown has started.&lt;br /&gt;
&lt;br /&gt;
''' SciNet/Niagara Downtime Announcement, May 6-7, 2020'''&lt;br /&gt;
&lt;br /&gt;
All resources at SciNet will undergo a two-day maintenance shutdown on May 6th and 7th 2020, starting at 7 am EDT on Wednesday May 6th.  There will be no access to any of the SciNet systems (Niagara, Mist, HPSS, Teach cluster, or the file systems) or systems hosted at the SciNet data centre.  We expect to be able to bring the systems back online the evening of May 7th.&lt;br /&gt;
&lt;br /&gt;
'''May 4, 2020, 7:51 AM:''' A power glitch this morning caused compute nodes to be rebooted: jobs running at the time may have failed; users are asked to resubmit these jobs.&lt;br /&gt;
&lt;br /&gt;
'''May 3, 2020, 8:20 AM:''' A power glitch this morning caused all compute nodes to be rebooted: jobs running at the time may have failed; users are asked to resubmit these jobs.&lt;br /&gt;
&lt;br /&gt;
'''April 28, 2020, 7:20 AM:''' A power glitch this morning caused all compute nodes to be rebooted: jobs running at the time have failed; users are asked to resubmit these jobs.&lt;br /&gt;
 &lt;br /&gt;
'''April 20, 2020: Security Incident at Cedar; implications for Niagara users'''&lt;br /&gt;
&lt;br /&gt;
Last week, it became evident that the Cedar GP cluster had been&lt;br /&gt;
comprimised for several weeks.  The passwords of at least two&lt;br /&gt;
Compute Canada users were known to the attackers. One of these was&lt;br /&gt;
used to escalate privileges on Cedar, as explained on&lt;br /&gt;
https://status.computecanada.ca/view_incident?incident=423.&lt;br /&gt;
&lt;br /&gt;
These accounts were used to login to Niagara as well, but Niagara&lt;br /&gt;
did not have the same security loophole as Cedar (which has been&lt;br /&gt;
fixed), and no further escalation was observed on Niagara.&lt;br /&gt;
&lt;br /&gt;
Reassuring as that may sound, it is not known how the passwords of&lt;br /&gt;
the two user accounts were obtained. Given this uncertainty, the&lt;br /&gt;
SciNet team *strongly* recommends that you change your password on&lt;br /&gt;
https://ccdb.computecanada.ca/security/change_password, and remove&lt;br /&gt;
any SSH keys and regenerate new ones (see&lt;br /&gt;
https://docs.scinet.utoronto.ca/index.php/SSH_keys).&lt;br /&gt;
&lt;br /&gt;
''' Tue 30 Mar 2020 14:55:14 EDT'''  Burst Buffer available again.&lt;br /&gt;
&lt;br /&gt;
''' Fri Mar 27 15:29:00 EDT 2020:''' SciNet systems are back up. Only the Burst Buffer remains offline, its maintenance is expected to be finished early next week.&lt;br /&gt;
&lt;br /&gt;
''' Thu Mar 26 23:05:00 EDT 2020:'''  Some aspects of the maintenance took longer than expected. The systems will not be back up until some time tomorrow, Friday March 27, 2020.  &lt;br /&gt;
&lt;br /&gt;
''' Wed Mar 25 7:00:00 EDT 2020:'''  SciNet/Niagara downtime started.&lt;br /&gt;
&lt;br /&gt;
''' Mon Mar 23 18:45:10 EDT 2020:'''  File system issues were resolved.&lt;br /&gt;
&lt;br /&gt;
''' Mon Mar 23 18:01:19 EDT 2020:''' There is currently an issue with the main Niagara filesystems. This effects all systems, all jobs have been killed. The issue is being investigated. &lt;br /&gt;
&lt;br /&gt;
''' Fri Mar 20 13:15:33 EDT 2020: ''' There was a power glitch at the datacentre at 8:50 AM, which resulted in jobs getting killed.  Please resubmit failed jobs. &lt;br /&gt;
&lt;br /&gt;
''' COVID-19 Impact on SciNet Operations, March 18, 2020'''&lt;br /&gt;
&lt;br /&gt;
Although the University of Toronto is closing of some of its&lt;br /&gt;
research operations on Friday March 20 at 5 pm EDT, this does not&lt;br /&gt;
affect the SciNet systems (such as Niagara, Mist, and HPSS), which&lt;br /&gt;
will remain operational.&lt;br /&gt;
&lt;br /&gt;
''' SciNet/Niagara Downtime Announcement, March 25-26, 2020'''&lt;br /&gt;
&lt;br /&gt;
All resources at SciNet will undergo a two-day maintenance shutdown on March 25th and 26th 2020, starting at 7 am EDT on Wednesday March 25th.  There will be no access to any of the SciNet systems (Niagara, Mist, HPSS, Teach cluster, or the file systems) during this time.&lt;br /&gt;
&lt;br /&gt;
This shutdown is necessary to finish the expansion of the Niagara cluster and its storage system.&lt;br /&gt;
&lt;br /&gt;
We expect to be able to bring the systems back online the evening of March 26th.&lt;br /&gt;
&lt;br /&gt;
''' March 9, 2020, 11:24 PM:''' HPSS services are temporarily suspended for emergency maintenance.&lt;br /&gt;
&lt;br /&gt;
''' March 7, 2020, 10:15 PM:''' File system issues have been cleared.&lt;br /&gt;
&lt;br /&gt;
''' March 6, 2020, 7:30 PM:''' File system issues; we are investigating&lt;br /&gt;
&lt;br /&gt;
''' March 2, 2020, 1:30 PM:''' For the extension of Niagara, the operating system on all Niagara nodes has been upgraded&lt;br /&gt;
from CentOS 7.4 to 7.6.  This required all&lt;br /&gt;
nodes to be rebooted. Running compute jobs are allowed to finish&lt;br /&gt;
before the compute node gets rebooted. Login nodes have all been rebooted, as have the datamover nodes and the jupyterhub service.&lt;br /&gt;
&lt;br /&gt;
''' Feb 24, 2020, 1:30PM: ''' The [[Mist]] login node got rebooted.  It is back, but we are still monitoring the situation.&lt;br /&gt;
&lt;br /&gt;
''' Feb 12, 2020, 11:00AM: ''' The [[Mist]] GPU cluster now available to users.&lt;br /&gt;
&lt;br /&gt;
''' Feb 11, 2020, 2:00PM: ''' The Niagara compute nodes were accidentally rebooted, killing all running jobs.&lt;br /&gt;
&lt;br /&gt;
''' Feb 10, 2020, 19:00PM: ''' HPSS is back to normal.&lt;br /&gt;
&lt;br /&gt;
''' Jan 30, 2020, 12:01PM: ''' We are having an issue with HPSS, in which the disk-cache is full. We put a reservation on the whole system (Globus, plus archive and vfs queues), until it has had a chance to clear some space on the cache.&lt;br /&gt;
&lt;br /&gt;
''' Jan 21, 2020, 4:05PM: '''   The was a partial power outage the took down a large amount of the compute nodes.  If your job died during this period please resubmit.  &lt;br /&gt;
&lt;br /&gt;
'''Jan 13, 2020, 7:35 PM:''' Maintenance finished.&lt;br /&gt;
&lt;br /&gt;
'''Jan 13, 2020, 8:20 AM:''' The announced maintenance downtime started (see below).&lt;br /&gt;
&lt;br /&gt;
'''Jan 9 2020, 11:30 AM:''' External ssh connectivity restored, issue related to the university network.&lt;br /&gt;
&lt;br /&gt;
'''Jan 9 2020, 9:24 AM:''' We received reports of users having trouble connecting into the SciNet data centre; we're investigating.  Systems are up and running and jobs are fine.&amp;lt;p&amp;gt;&lt;br /&gt;
As a work around, in the meantime, it appears to be possible to log into graham, cedar or beluga, and then ssh to niagara.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Downtime announcement:'''&lt;br /&gt;
To prepare for the upcoming expansion of Niagara, there will be a&lt;br /&gt;
one-day maintenance shutdown on '''January 13th 2020, starting at 8 am&lt;br /&gt;
EST'''.  There will be no access to Niagara, Mist, HPSS or teach, nor&lt;br /&gt;
to their file systems during this time.&lt;br /&gt;
&lt;br /&gt;
2019&lt;br /&gt;
&lt;br /&gt;
'''December 13, 9:00 AM EST:''' Issues resolved.&lt;br /&gt;
&lt;br /&gt;
'''December 13, 8:20 AM EST:''' Overnight issue is now preventing logins to Niagara and other services. Possibly a file system issue, we are investigating.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt; '''Fri, Nov 15 2019, 11:00 PM (EST)'''  Niagara and most of the main systems are now available. &lt;br /&gt;
&amp;lt;/p&amp;gt;&amp;lt;p&amp;gt; '''Fri, Nov 15 2019, 7:50 PM (EST)'''  SOSCIP GPU cluster is up and accessible.  Work on the other systems continues.&lt;br /&gt;
&amp;lt;/p&amp;gt;&amp;lt;p&amp;gt; '''Fri, Nov 15 2019, 5:00 PM (EST)'''  Infrastructure maintenance done, upgrades still in process.&lt;br /&gt;
&amp;lt;/p&amp;gt;&amp;lt;p&amp;gt;&lt;br /&gt;
'''Fri, Nov 15 2019, 7:00 AM (EST)'''  Maintenance shutdown of the SciNet data centre has started.  Note: scratch purging has been postponed until Nov 17.&amp;lt;br/&amp;gt; &lt;br /&gt;
&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&lt;br /&gt;
'''Announcement:''' &lt;br /&gt;
The SciNet datacentre will undergo a maintenance shutdown on&lt;br /&gt;
Friday November 15th 2019, from 7 am to 11 pm (EST), with no access&lt;br /&gt;
to any of the SciNet systems (Niagara, P8, SGC, HPSS, Teach cluster,&lt;br /&gt;
or the filesystems) during that time. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
'''Sat, Nov 2 2019, 1:30 PM (update):'''  Chiller has been fixed, all systems are operational.    &lt;br /&gt;
&amp;lt;/p&amp;gt;&lt;br /&gt;
'''Fri, Nov 1 2019, 4:30 PM (update):'''  We are operating in free cooling so have brought up about 1/2 of the Niagara compute nodes to reduce the cooling load.  Access, storage, and other systems should now be available.   &lt;br /&gt;
&lt;br /&gt;
'''Fri, Nov 1 2019, 12:05 PM (update):''' A power module in the chiller has failed and needs to be replaced.   We should be able to operate in free cooling if the temperature stays cold enough, but we may not be able to run all systems. No eta yet on when users will be able to log back in. &lt;br /&gt;
&lt;br /&gt;
'''Fri, Nov 1 2019, 9:15 AM (update):''' There was a automated shutdown because of rising temperatures, causing all systems to go down. We are investigating, check here for updates.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;'''Fri, Nov 1 2019, 8:16 AM:''' Unexpected data centre issue: Check here for updates.&lt;br /&gt;
&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
''' Thu 1 Aug 2019 5:00:00 PM ''' Systems are up and operational.   &lt;br /&gt;
&lt;br /&gt;
'''Thu 1 Aug 2019 7:00:00 AM: ''' Scheduled Downtime Maintenance of the SciNet Datacenter.  All systems will be down and unavailable starting 7am until the evening. &lt;br /&gt;
&lt;br /&gt;
'''Fri 26 Jul 2019, 16:02:26 EDT:''' There was an issue with the Burst Buffer at around 3PM, and it was recently solved. BB is OK again.&lt;br /&gt;
&lt;br /&gt;
''' Sun 30 Jun 2019 ''' The '''SOSCIP BGQ''' and '''P7''' systems were decommissioned on '''June 30th, 2019'''.  The BGQdev front end node and storage are still available.  &lt;br /&gt;
&lt;br /&gt;
'''Wed 19 Jun 2018, 1:20:00 PM:''' The BGQ is back online.&lt;br /&gt;
&lt;br /&gt;
'''Wed 19 Jun 2018, 10:00:00 AM:''' The BGQ is still down, the SOSCIP GPU nodes should be back up. &lt;br /&gt;
&lt;br /&gt;
'''Wed 19 Jun 2018, 1:40:00 AM:''' There was an issue with the SOSCIP BGQ and GPU Cluster last night about 1:42am, probably a power fluctuation that took it down.  &lt;br /&gt;
&lt;br /&gt;
'''Wed 12 Jun 2019, 3:30 AM - 7:40 AM''' Intermittent system issues on Niagara's project and scratch as the file number limit was reached. We increased the number of files allowed in total on the file system. &lt;br /&gt;
&lt;br /&gt;
'''Thu 30 May 2019, 11:00:00 PM:'''&lt;br /&gt;
The maintenance downtime of SciNet's data center has finished, and systems are being brought online now.  You can check the progress here. Some systems might not be available until Friday morning.&amp;lt;br/&amp;gt;&lt;br /&gt;
Some action on the part of users will be required when they first connect again to a Niagara login nodes or datamovers.  This is due to the security upgrade of the Niagara cluster, which is now in line with currently accepted best practices.&amp;lt;br/&amp;gt;&lt;br /&gt;
The details of the required actions can be found on the [[SSH Changes in May 2019]] wiki page.&lt;br /&gt;
&lt;br /&gt;
'''Wed 29-30 May 2019''' The SciNet datacentre will undergo a two-day maintenance shutdown, starting at 7 am EDT on Wednesday May 29th.  There will be no access to any of the SciNet systems (Niagara, P7, P8, BGQ, SGC, HPSS, Teach cluster, or the file systems) during this time.&lt;br /&gt;
&lt;br /&gt;
'''SCHEDULED SHUTDOWN''': &lt;br /&gt;
&lt;br /&gt;
Please be advised that on '''Wednesday May 29th through Thursday May 30th''', the SciNet datacentre will undergo a two-day maintenance shutdown, starting at 7 am EDT on Wednesday May 29th.  There will be no access to any of the SciNet systems (Niagara, P7, P8, BGQ, SGC, HPSS, Teach cluster, or the file systems) during this time.&lt;br /&gt;
&lt;br /&gt;
This is necessary to finish the installation of an emergency power generator, to perform the annual cooling tower maintenance, and to enhance login security.&lt;br /&gt;
&lt;br /&gt;
We expect to be able to bring the systems back online the evening of May 30th.  Due to the enhanced login security, the ssh applications of users will need to update their known host list. More detailed information on this procedure will be sent shortly before the systems are back online.&lt;br /&gt;
&lt;br /&gt;
'''Fri 5 Apr 2019:''' Software updates on Niagara: The default CCEnv software stack now uses avx512 on Niagara, and there is now a NiaEnv/2019b stack (&amp;quot;epoch&amp;quot;). &lt;br /&gt;
&lt;br /&gt;
'''Thu 4 Apr 2019:''' The 2019 compute and storage allocations have taken effect on Niagara.&lt;br /&gt;
&lt;br /&gt;
'''NOTE''':  There is scheduled network maintenance for '''Friday April 26th 12am-8am''' on the Scinet datacenter external network connection.   This will not affect internal connections and running jobs however remote connections may see interruptions during this period.&lt;br /&gt;
&lt;br /&gt;
'''Wed 24 Apr 2019 14:14 EDT:''' HPSS is back on service. Library and robot arm maintenance finished.&lt;br /&gt;
&lt;br /&gt;
'''Wed 24 Apr 2019 08:35 EDT:''' HPSS out of service this morning for library and robot arm maintenance.&lt;br /&gt;
&lt;br /&gt;
'''Fri 19 Apr 2019 17:40 EDT:''' HPSS robot arm has been released and is back to normal operations.&lt;br /&gt;
&lt;br /&gt;
'''Fri 19 Apr 2019 14:00 EDT:''' problems with HPPS library robot have been detected.&lt;br /&gt;
&lt;br /&gt;
'''Wed 17 Apr 2019 15:35 EDT:''' Network connection is back.&lt;br /&gt;
&lt;br /&gt;
'''Wed 17 Apr 2019 15:12 EDT:''' Network connection down.  Investigating.&lt;br /&gt;
&lt;br /&gt;
'''Tue 9 Apr 2019 22:24:14 EDT:'''  Network connection restored.&lt;br /&gt;
&lt;br /&gt;
'''Tue 9 Apr 2019, 15:20:''' Network connection down.  Investigating.&lt;br /&gt;
&lt;br /&gt;
'''Fri 5 Apr 2019:''' Planned, short outage in connectivity to the SciNet datacentre from 7:30 am to 8:55 am EST for maintenance of the network.  This outage will not affect running or queued jobs. It may be necessary to reboot the login nodes at some point tomorrow, which could result in a short interruption of connectivity, but which will have no effect on running or queued jobs.&lt;br /&gt;
&lt;br /&gt;
'''April 4, 2019:'''  The 2019 compute and storage allocations will take effect on Niagara. Running jobs will not be affected by this change and will run their course.  Queued jobs' priorities will be updated to reflect the new fairshare values later in the day.  The queue should fully reflect the new fairshare values in about 24 hours.   &lt;br /&gt;
&lt;br /&gt;
It may be necessary to reboot the login nodes at some point tomorrow, which could result in a short interruption of connectivity, but which will have no effect on running or queued jobs.&lt;br /&gt;
&lt;br /&gt;
There will be updates to the software stack on this day as well.&lt;br /&gt;
&lt;br /&gt;
'''March 25, 3:05 PM EST:'''  Most systems back online, other services should be back shortly. &lt;br /&gt;
&lt;br /&gt;
'''March 25, 12:05 PM EST:''' Power is back at the datacentre, but it is not yet known when all systems will be back up.  Keep checking here for updates.&lt;br /&gt;
&lt;br /&gt;
'''March 25, 11:27 AM EST:''' A power outage in the datacentre occured and caused all services to go down.  Check here for updates.&lt;br /&gt;
&lt;br /&gt;
'''Thu Mar 21 10:37:28 EDT 2019:''' HPSS is back in service&lt;br /&gt;
&lt;br /&gt;
HPSS out of service on '''Tue, Mar/19 at 9AM''', for tape library expansion and relocation. It's possible the downtime will extend to Wed, Mar/20.&lt;br /&gt;
&lt;br /&gt;
'''January 21, 4:00 PM''': HPSS is back in service. Thank you for your patience.&lt;br /&gt;
&lt;br /&gt;
'''January 18, 5:00 PM''': We did practically all of the HPSS upgrades (software/hardware), however the main client node - archive02 - is presenting an issue we just couldn't resolve yet. We will try to resume work over the weekend with cool heads, or on Monday. Sorry, but this is an unforeseen delay. Jobs on the queue we'll remain there, and we'll delay the scratch purging by 1 week.&lt;br /&gt;
&lt;br /&gt;
'''January 16, 11:00 PM''': HPSS is being upgraded, as announced.&lt;br /&gt;
&lt;br /&gt;
'''January 16, 8:00 PM''': System are coming back up and should be accessible for users now.&lt;br /&gt;
&lt;br /&gt;
'''January 15, 8:00 AM''': Data centre downtime in effect.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;font color=red&amp;gt;&amp;lt;b&amp;gt;Downtime Announcement for January 15 and 16, 2019&amp;lt;/b&amp;gt;&amp;lt;/font&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
The SciNet datacentre will need to undergo a two-day maintenance shutdown in order to perform electrical work, repairs and maintenance.  The electrical work is in preparation for the upcoming installation of an emergency power generator and a larger UPS, which will result in increased resilience to power glitches and outages.  The shutdown is scheduled to start on '''Tuesday January 15, 2019, at 7 am''' and will last until '''Wednesday 16, 2019''', some time in the evening. There will be no access to any of the SciNet systems (Niagara, P7, P8, BGQ, SGC, HPSS, Teach cluster, or the filesystems) during this time.&lt;br /&gt;
Check back here for up-to-date information on the status of the systems.&lt;br /&gt;
&lt;br /&gt;
Note: this downtime was originally scheduled for Dec. 18, 2018, but has been postponed and combined with the annual maintenance downtime.&lt;br /&gt;
&lt;br /&gt;
'''December 24, 2018, 11:35 AM EST:''' Most systems are operational again. If you had compute jobs running yesterday at around 3:30PM, they likely crashed - please check them and resubmit if needed.&lt;br /&gt;
&lt;br /&gt;
'''December 24, 2018, 10:40 AM EST:''' Repairs have been made, and the file systems are starting to be mounted on the cluster. &lt;br /&gt;
&lt;br /&gt;
'''December 23, 2018, 3:38 PM EST:''' Issues with the file systems (home, scratch and project). We are investigating, it looks like a hardware issue that we are trying to work around. Note that the absence of /home means you cannot log in with ssh keys. All compute jobs crashed around 3:30 PM EST on Dec 23. Once the system is properly up again, please resubmit your jobs.  Unfortunately, at this time of year, it is not possible to give an estimate on when the system will be operational again.&lt;br /&gt;
&lt;br /&gt;
'''Tue Nov 22 14:20:00 EDT 2018''': &amp;lt;font color=green&amp;gt;HPSS back in service&amp;lt;/font&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Tue Nov 22 08:55:00 EDT 2018''': &amp;lt;font color=red&amp;gt;HPSS offline for scheduled maintenance&amp;lt;/font&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Tue Nov 20 16:30:00 EDT 2018''':  HPSS offline on Thursday 9AM for installation of new LTO8 drives in the tape library.&lt;br /&gt;
&lt;br /&gt;
'''Tue Oct  9 12:16:00 EDT 2018''':  BGQ compute nodes are up.  &lt;br /&gt;
&lt;br /&gt;
'''Sun Oct  7 20:24:26 EDT 2018''':  SGC and BGQ front end are available,  BGQ compute nodes down related to a cooling issue.  &lt;br /&gt;
&lt;br /&gt;
'''Sat Oct  6 23:16:44 EDT 2018''':  There were some problems bringing up SGC &amp;amp; BGQ, they will remain offline for now.&lt;br /&gt;
&lt;br /&gt;
'''Sat Oct  6 18:36:35 EDT 2018''':  Electrical work finished, power restored. Systems are coming online.&lt;br /&gt;
&lt;br /&gt;
'''July 18, 2018:''' login.scinet.utoronto.ca is now disabled, GPC $SCRATCH and $HOME are decommissioned.&lt;br /&gt;
&lt;br /&gt;
'''July 12, 2018:''' There was a short power interruption around 10:30 am which caused most of the systems (Niagara, SGC, BGQ) to reboot and any running jobs to fail. &lt;br /&gt;
&lt;br /&gt;
'''July 11, 2018:''' P7's moved to BGQ filesystem, P8's moved to Niagara filesystem.&lt;br /&gt;
&lt;br /&gt;
'''May 24, 2018, 9:25 PM EST:''' The data center is up, and all systems are operational again.&lt;br /&gt;
&lt;br /&gt;
'''May 24, 2018, 7:00 AM EST:''' The data centre is under annual maintenance. All systems are offline. Systems are expected to be back late afternoon today; check for updates on this page.&lt;br /&gt;
&lt;br /&gt;
'''May 18, 2018:''' Announcement: Annual scheduled maintenance downtime: Thursday May 24, starting 7:00 AM&lt;br /&gt;
&lt;br /&gt;
'''May 16, 2018:''' Cooling  restored, systems online&lt;br /&gt;
&lt;br /&gt;
'''May 16, 2018:''' Cooling issue at datacentre again, all systems down&lt;br /&gt;
&lt;br /&gt;
'''May 15, 2018:''' Cooling restored, systems coming online&lt;br /&gt;
&lt;br /&gt;
'''May 15, 2018''' Cooling issue at datacentre, all systems down&lt;br /&gt;
&lt;br /&gt;
'''May 4, 2018:''' [[HPSS]] is now operational on Niagara.&lt;br /&gt;
&lt;br /&gt;
'''May 3, 2018:''' [[Burst Buffer]] is available upon request.&lt;br /&gt;
&lt;br /&gt;
'''May 3, 2018:''' The [https://docs.computecanada.ca/wiki/Globus Globus] endpoint for Niagara is available: computecanada#niagara.&lt;br /&gt;
&lt;br /&gt;
'''May 1, 2018:''' System status moved he here.&lt;br /&gt;
&lt;br /&gt;
'''Apr 23, 2018:''' GPC-compute is decommissioned, GPC-storage available until 30 May 2018.&lt;br /&gt;
&lt;br /&gt;
'''April 10, 2018:''' Niagara commissioned.&lt;/div&gt;</summary>
		<author><name>Rzon</name></author>
	</entry>
	<entry>
		<id>https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7544</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7544"/>
		<updated>2026-02-17T21:45:58Z</updated>

		<summary type="html">&lt;p&gt;Rzon: Undo revision 7541 by Rzon (talk)&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;__NOTOC__&lt;br /&gt;
{| style=&amp;quot;border-spacing:10px; width: 95%&amp;quot;&lt;br /&gt;
| style=&amp;quot;padding:1em; padding-top:.1em; border:2px solid #0645ad; background-color:#f6f6f6; border-radius:7px&amp;quot;|&lt;br /&gt;
&lt;br /&gt;
==System Status==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- Use &amp;quot;Up&amp;quot;, &amp;quot;Partial&amp;quot; or &amp;quot;Down&amp;quot;; these are templates. --&amp;gt;&lt;br /&gt;
{|style=&amp;quot;width:100%&amp;quot; &lt;br /&gt;
|{{Down3 | Trillium|https://docs.alliancecan.ca/wiki/Trillium_Quickstart}}&lt;br /&gt;
|{{Up3 | OnDemand|https://docs.alliancecan.ca/wiki/Trillium_Open_OnDemand_Quickstart}}&lt;br /&gt;
|{{Up | Globus |Globus}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Up | HPSS|HPSS}}&lt;br /&gt;
|{{Up | Balam|Balam}}&lt;br /&gt;
|{{Partial | S4H | S4H}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Up | Teach|Teach}}&lt;br /&gt;
|{{Up3 | File system|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Storage}}&lt;br /&gt;
|{{Up3 | External Network|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Logging_in}} &lt;br /&gt;
|}&lt;br /&gt;
'''Tue Feb 17, 2026, 8:40 am:''' Power outage at the data centre.  Cooling issues have developed as a result.  Major systems (Trillium, S4H) are expected to be down until sometime Thursday. Login nodes and file systems will remain accessible.&lt;br /&gt;
&lt;br /&gt;
'''Mon Feb 16, 2026, 8:40 pm:''' Electricity is unstable in the data centre area due to severe snowfall.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 1:40 pm:''' All services are operational again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 12:00 pm:''' The Trillium and Open OnDemand compute nodes are operational again. We are still working on bringing Balam, Neptune and S4H nodes up again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 10:00 am:''' There was a power glitch at the data centre overnight. The login nodes are accessible but the compute nodes are down.  &lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 16, 2026, 11:00 pm:''' HPSS is back online, and accessible via alliancecan#hpss Globus endpoint. &lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 15, 2026, 10:00 pm:''' HPSS will undergo maintenance on Friday morning, Jan/16/2025, , including alliancecan#hpss Globus endpoint &lt;br /&gt;
&lt;br /&gt;
'''Tue Jan 6, 2026, 10:15 am:''' OnDemand has been fixed and is working again.&lt;br /&gt;
&lt;br /&gt;
'''Mon Jan 5, 2026, 9:00 pm:''' The authentication mechanism of OnDemand is not working.&lt;br /&gt;
&lt;br /&gt;
'''Wed Dec 31, 2025, 12:40 pm:''' We believe the problem has now been resolved.  Please let us know if you still experience login problems or aborted jobs.&lt;br /&gt;
&lt;br /&gt;
'''Tue Dec 30, 2025, 2:10 pm:''' We are experiencing problems with authentication, resulting in failed logins, OOD errors, and aborted jobs (with &amp;quot;prolog error&amp;quot;).  Please bear with us, as we are very short-staffed during the holiday break.  We will post updates here.&lt;br /&gt;
&lt;br /&gt;
'''Tue Dec 3, 2025, 11:30 am:''' Open OnDemand is fully operational again.&lt;br /&gt;
&lt;br /&gt;
'''Sat Nov 29, 2025, 00:40 am:''' There has been a problem with the water chiller. Some systems are offline.&lt;br /&gt;
&lt;br /&gt;
'''Wed Nov 5, 2025, 12:55 pm:''' Balam is back online.&lt;br /&gt;
&lt;br /&gt;
'''Wed Nov 5, 2025, 10:00 am:''' Open OnDemand is back online.&lt;br /&gt;
&lt;br /&gt;
'''Tue Nov 4, 2025, 11:00 pm:''' Most of the work is done, data movers, Globus, and HPSS are back online. Remaining services will be worked on tomorrow.&lt;br /&gt;
&lt;br /&gt;
'''Tue Nov 4, 2025, 8:30 am:''' Scheduled network maintenance. Trillium cluster is *not* affected.&lt;br /&gt;
&lt;br /&gt;
'''Tue Oct 21, 2025, 17:30 am:''' Balam maintenance finished.&lt;br /&gt;
&lt;br /&gt;
'''Tue Oct 21, 2025, 7:00 am:''' Balam maintenance day.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 15, 2025, 3:55 pm:''' Trillium inbound connections through trillium.alliancecan.ca or trillium.scinet.utoronto.ca are working again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 15, 2025, 3:05 pm:''' Trillium is experiencing external network issues for both incoming traffic. Please try: ssh USERNAME@tri-login01.scinet.utoronto.ca in the meantime.&lt;br /&gt;
 &lt;br /&gt;
'''Thu Oct 06, 2025, 8:00 pm:''' HPSS is fully functional. You may submit archive jobs from trillium login nodes, datamovers and robots.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 03, 2025, 6:30 pm:''' HPSS is back online, and already accessible via alliancecan#hpss Globus endpoint. Directory tree now follows the other Alliance clusters. We're still working on job submission via Slurm&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 01, 2025, 0:00 am:''' Niagara compute nodes are now unavailable for regular users. The login nodes will remain available for a while to allow a few last data transfers, although transfers from the Niagara file systems to Trillium are best done on nia-dm1.scinet.utoronto.ca.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 01, 2025, 9:30 am:''' HPSS is down for scheduled maintenance, including alliancecan#hpss Globus endpoint&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--  When removing system status entries, please archive them to: --&amp;gt;&lt;br /&gt;
[[Previous messages]]&lt;br /&gt;
&lt;br /&gt;
{|style=&amp;quot;border-spacing: 10px;width: 100%&amp;quot;&lt;br /&gt;
|valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== QuickStart Guides ==&lt;br /&gt;
* [https://docs.alliancecan.ca/wiki/Trillium_Quickstart Trillium Quickstart]&lt;br /&gt;
* [[Niagara Quickstart]]&lt;br /&gt;
* [[HPSS | HPSS archival storage]]&lt;br /&gt;
* [[Teach|Teach cluster]]&lt;br /&gt;
* [[FAQ | FAQ (frequently asked questions)]]&lt;br /&gt;
* [[Acknowledging SciNet]]&lt;br /&gt;
| valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== Tutorials, Manuals, etc. ==&lt;br /&gt;
* [https://education.scinet.utoronto.ca SciNet education material]&lt;br /&gt;
* [https://www.youtube.com/c/SciNetHPCattheUniversityofToronto SciNet's YouTube channel]&lt;br /&gt;
* [[Modules specific to Niagara|Software Modules specific to Niagara]] &lt;br /&gt;
* [[Modules for Mist]] &lt;br /&gt;
* [[Commercial software]]&lt;br /&gt;
* [[Burst Buffer]]&lt;br /&gt;
* [[SSH#SSH Keys|SSH keys]]&lt;br /&gt;
* [[SSH Tunneling]]&lt;br /&gt;
* [[Visualization]]&lt;br /&gt;
* [[Running Serial Jobs on Niagara]]&lt;br /&gt;
* [[Jupyter Hub]]&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Rzon</name></author>
	</entry>
	<entry>
		<id>https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7541</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7541"/>
		<updated>2026-02-17T21:39:32Z</updated>

		<summary type="html">&lt;p&gt;Rzon: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;__NOTOC__&lt;br /&gt;
{| style=&amp;quot;border-spacing:10px; width: 95%&amp;quot;&lt;br /&gt;
| style=&amp;quot;padding:1em; padding-top:.1em; border:2px solid #0645ad; background-color:#f6f6f6; border-radius:7px&amp;quot;|&lt;br /&gt;
&lt;br /&gt;
==System Status==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- Use &amp;quot;Up&amp;quot;, &amp;quot;Partial&amp;quot; or &amp;quot;Down&amp;quot;; these are templates. --&amp;gt;&lt;br /&gt;
{|style=&amp;quot;width:100%&amp;quot; &lt;br /&gt;
|{{Down3 | Trillium|https://docs.alliancecan.ca/wiki/Trillium_Quickstart}}&lt;br /&gt;
|{{Down3 | OnDemand|https://docs.alliancecan.ca/wiki/Trillium_Open_OnDemand_Quickstart}}&lt;br /&gt;
|{{Down | Globus |Globus}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Down | HPSS|HPSS}}&lt;br /&gt;
|{{Down | Balam|Balam}}&lt;br /&gt;
|{{Down | S4H | S4H}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Down | Teach|Teach}}&lt;br /&gt;
|{{Down3 | File system|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Storage}}&lt;br /&gt;
|{{Down3 | External Network|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Logging_in}} &lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Tue Feb 17, 2026, 4:35 pm:''' SciNet data centre went down. We are investigating.&lt;br /&gt;
&lt;br /&gt;
'''Tue Feb 17, 2026, 8:40 am:''' Power outage at the data centre.  Cooling issues have developed as a result.  Major systems (Trillium, S4H) are expected to be down until sometime Thursday. Login nodes and file systems will remain accessible.&lt;br /&gt;
&lt;br /&gt;
'''Mon Feb 16, 2026, 8:40 pm:''' Electricity is unstable in the data centre area due to severe snowfall.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 1:40 pm:''' All services are operational again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 12:00 pm:''' The Trillium and Open OnDemand compute nodes are operational again. We are still working on bringing Balam, Neptune and S4H nodes up again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 10:00 am:''' There was a power glitch at the data centre overnight. The login nodes are accessible but the compute nodes are down.  &lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 16, 2026, 11:00 pm:''' HPSS is back online, and accessible via alliancecan#hpss Globus endpoint. &lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 15, 2026, 10:00 pm:''' HPSS will undergo maintenance on Friday morning, Jan/16/2025, , including alliancecan#hpss Globus endpoint &lt;br /&gt;
&lt;br /&gt;
'''Tue Jan 6, 2026, 10:15 am:''' OnDemand has been fixed and is working again.&lt;br /&gt;
&lt;br /&gt;
'''Mon Jan 5, 2026, 9:00 pm:''' The authentication mechanism of OnDemand is not working.&lt;br /&gt;
&lt;br /&gt;
'''Wed Dec 31, 2025, 12:40 pm:''' We believe the problem has now been resolved.  Please let us know if you still experience login problems or aborted jobs.&lt;br /&gt;
&lt;br /&gt;
'''Tue Dec 30, 2025, 2:10 pm:''' We are experiencing problems with authentication, resulting in failed logins, OOD errors, and aborted jobs (with &amp;quot;prolog error&amp;quot;).  Please bear with us, as we are very short-staffed during the holiday break.  We will post updates here.&lt;br /&gt;
&lt;br /&gt;
'''Tue Dec 3, 2025, 11:30 am:''' Open OnDemand is fully operational again.&lt;br /&gt;
&lt;br /&gt;
'''Sat Nov 29, 2025, 00:40 am:''' There has been a problem with the water chiller. Some systems are offline.&lt;br /&gt;
&lt;br /&gt;
'''Wed Nov 5, 2025, 12:55 pm:''' Balam is back online.&lt;br /&gt;
&lt;br /&gt;
'''Wed Nov 5, 2025, 10:00 am:''' Open OnDemand is back online.&lt;br /&gt;
&lt;br /&gt;
'''Tue Nov 4, 2025, 11:00 pm:''' Most of the work is done, data movers, Globus, and HPSS are back online. Remaining services will be worked on tomorrow.&lt;br /&gt;
&lt;br /&gt;
'''Tue Nov 4, 2025, 8:30 am:''' Scheduled network maintenance. Trillium cluster is *not* affected.&lt;br /&gt;
&lt;br /&gt;
'''Tue Oct 21, 2025, 17:30 am:''' Balam maintenance finished.&lt;br /&gt;
&lt;br /&gt;
'''Tue Oct 21, 2025, 7:00 am:''' Balam maintenance day.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 15, 2025, 3:55 pm:''' Trillium inbound connections through trillium.alliancecan.ca or trillium.scinet.utoronto.ca are working again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 15, 2025, 3:05 pm:''' Trillium is experiencing external network issues for both incoming traffic. Please try: ssh USERNAME@tri-login01.scinet.utoronto.ca in the meantime.&lt;br /&gt;
 &lt;br /&gt;
'''Thu Oct 06, 2025, 8:00 pm:''' HPSS is fully functional. You may submit archive jobs from trillium login nodes, datamovers and robots.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 03, 2025, 6:30 pm:''' HPSS is back online, and already accessible via alliancecan#hpss Globus endpoint. Directory tree now follows the other Alliance clusters. We're still working on job submission via Slurm&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 01, 2025, 0:00 am:''' Niagara compute nodes are now unavailable for regular users. The login nodes will remain available for a while to allow a few last data transfers, although transfers from the Niagara file systems to Trillium are best done on nia-dm1.scinet.utoronto.ca.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 01, 2025, 9:30 am:''' HPSS is down for scheduled maintenance, including alliancecan#hpss Globus endpoint&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--  When removing system status entries, please archive them to: --&amp;gt;&lt;br /&gt;
[[Previous messages]]&lt;br /&gt;
&lt;br /&gt;
{|style=&amp;quot;border-spacing: 10px;width: 100%&amp;quot;&lt;br /&gt;
|valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== QuickStart Guides ==&lt;br /&gt;
* [https://docs.alliancecan.ca/wiki/Trillium_Quickstart Trillium Quickstart]&lt;br /&gt;
* [[Niagara Quickstart]]&lt;br /&gt;
* [[HPSS | HPSS archival storage]]&lt;br /&gt;
* [[Teach|Teach cluster]]&lt;br /&gt;
* [[FAQ | FAQ (frequently asked questions)]]&lt;br /&gt;
* [[Acknowledging SciNet]]&lt;br /&gt;
| valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== Tutorials, Manuals, etc. ==&lt;br /&gt;
* [https://education.scinet.utoronto.ca SciNet education material]&lt;br /&gt;
* [https://www.youtube.com/c/SciNetHPCattheUniversityofToronto SciNet's YouTube channel]&lt;br /&gt;
* [[Modules specific to Niagara|Software Modules specific to Niagara]] &lt;br /&gt;
* [[Modules for Mist]] &lt;br /&gt;
* [[Commercial software]]&lt;br /&gt;
* [[Burst Buffer]]&lt;br /&gt;
* [[SSH#SSH Keys|SSH keys]]&lt;br /&gt;
* [[SSH Tunneling]]&lt;br /&gt;
* [[Visualization]]&lt;br /&gt;
* [[Running Serial Jobs on Niagara]]&lt;br /&gt;
* [[Jupyter Hub]]&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Rzon</name></author>
	</entry>
	<entry>
		<id>https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7538</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7538"/>
		<updated>2026-02-17T21:04:09Z</updated>

		<summary type="html">&lt;p&gt;Rzon: /* System Status */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;__NOTOC__&lt;br /&gt;
{| style=&amp;quot;border-spacing:10px; width: 95%&amp;quot;&lt;br /&gt;
| style=&amp;quot;padding:1em; padding-top:.1em; border:2px solid #0645ad; background-color:#f6f6f6; border-radius:7px&amp;quot;|&lt;br /&gt;
&lt;br /&gt;
==System Status==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- Use &amp;quot;Up&amp;quot;, &amp;quot;Partial&amp;quot; or &amp;quot;Down&amp;quot;; these are templates. --&amp;gt;&lt;br /&gt;
{|style=&amp;quot;width:100%&amp;quot; &lt;br /&gt;
|{{Down3 | Trillium|https://docs.alliancecan.ca/wiki/Trillium_Quickstart}}&lt;br /&gt;
|{{Up3 | OnDemand|https://docs.alliancecan.ca/wiki/Trillium_Open_OnDemand_Quickstart}}&lt;br /&gt;
|{{Up | Globus |Globus}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Up | HPSS|HPSS}}&lt;br /&gt;
|{{Up | Balam|Balam}}&lt;br /&gt;
|{{Partial | S4H | S4H}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Up | Teach|Teach}}&lt;br /&gt;
|{{Up3 | File system|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Storage}}&lt;br /&gt;
|{{Up3 | External Network|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Logging_in}} &lt;br /&gt;
|}&lt;br /&gt;
'''Tue Feb 17, 2026, 8:40 am:''' Power outage at the data centre.  Cooling issues have developed as a result.  Major systems (Trillium, S4H) are expected to be down until sometime Thursday. Login nodes and file systems will remain accessible.&lt;br /&gt;
&lt;br /&gt;
'''Mon Feb 16, 2026, 8:40 pm:''' Electricity is unstable in the data centre area due to severe snowfall.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 1:40 pm:''' All services are operational again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 12:00 pm:''' The Trillium and Open OnDemand compute nodes are operational again. We are still working on bringing Balam, Neptune and S4H nodes up again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 10:00 am:''' There was a power glitch at the data centre overnight. The login nodes are accessible but the compute nodes are down.  &lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 16, 2026, 11:00 pm:''' HPSS is back online, and accessible via alliancecan#hpss Globus endpoint. &lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 15, 2026, 10:00 pm:''' HPSS will undergo maintenance on Friday morning, Jan/16/2025, , including alliancecan#hpss Globus endpoint &lt;br /&gt;
&lt;br /&gt;
'''Tue Jan 6, 2026, 10:15 am:''' OnDemand has been fixed and is working again.&lt;br /&gt;
&lt;br /&gt;
'''Mon Jan 5, 2026, 9:00 pm:''' The authentication mechanism of OnDemand is not working.&lt;br /&gt;
&lt;br /&gt;
'''Wed Dec 31, 2025, 12:40 pm:''' We believe the problem has now been resolved.  Please let us know if you still experience login problems or aborted jobs.&lt;br /&gt;
&lt;br /&gt;
'''Tue Dec 30, 2025, 2:10 pm:''' We are experiencing problems with authentication, resulting in failed logins, OOD errors, and aborted jobs (with &amp;quot;prolog error&amp;quot;).  Please bear with us, as we are very short-staffed during the holiday break.  We will post updates here.&lt;br /&gt;
&lt;br /&gt;
'''Tue Dec 3, 2025, 11:30 am:''' Open OnDemand is fully operational again.&lt;br /&gt;
&lt;br /&gt;
'''Sat Nov 29, 2025, 00:40 am:''' There has been a problem with the water chiller. Some systems are offline.&lt;br /&gt;
&lt;br /&gt;
'''Wed Nov 5, 2025, 12:55 pm:''' Balam is back online.&lt;br /&gt;
&lt;br /&gt;
'''Wed Nov 5, 2025, 10:00 am:''' Open OnDemand is back online.&lt;br /&gt;
&lt;br /&gt;
'''Tue Nov 4, 2025, 11:00 pm:''' Most of the work is done, data movers, Globus, and HPSS are back online. Remaining services will be worked on tomorrow.&lt;br /&gt;
&lt;br /&gt;
'''Tue Nov 4, 2025, 8:30 am:''' Scheduled network maintenance. Trillium cluster is *not* affected.&lt;br /&gt;
&lt;br /&gt;
'''Tue Oct 21, 2025, 17:30 am:''' Balam maintenance finished.&lt;br /&gt;
&lt;br /&gt;
'''Tue Oct 21, 2025, 7:00 am:''' Balam maintenance day.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 15, 2025, 3:55 pm:''' Trillium inbound connections through trillium.alliancecan.ca or trillium.scinet.utoronto.ca are working again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 15, 2025, 3:05 pm:''' Trillium is experiencing external network issues for both incoming traffic. Please try: ssh USERNAME@tri-login01.scinet.utoronto.ca in the meantime.&lt;br /&gt;
 &lt;br /&gt;
'''Thu Oct 06, 2025, 8:00 pm:''' HPSS is fully functional. You may submit archive jobs from trillium login nodes, datamovers and robots.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 03, 2025, 6:30 pm:''' HPSS is back online, and already accessible via alliancecan#hpss Globus endpoint. Directory tree now follows the other Alliance clusters. We're still working on job submission via Slurm&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 01, 2025, 0:00 am:''' Niagara compute nodes are now unavailable for regular users. The login nodes will remain available for a while to allow a few last data transfers, although transfers from the Niagara file systems to Trillium are best done on nia-dm1.scinet.utoronto.ca.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 01, 2025, 9:30 am:''' HPSS is down for scheduled maintenance, including alliancecan#hpss Globus endpoint&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--  When removing system status entries, please archive them to: --&amp;gt;&lt;br /&gt;
[[Previous messages]]&lt;br /&gt;
&lt;br /&gt;
{|style=&amp;quot;border-spacing: 10px;width: 100%&amp;quot;&lt;br /&gt;
|valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== QuickStart Guides ==&lt;br /&gt;
* [https://docs.alliancecan.ca/wiki/Trillium_Quickstart Trillium Quickstart]&lt;br /&gt;
* [[Niagara Quickstart]]&lt;br /&gt;
* [[HPSS | HPSS archival storage]]&lt;br /&gt;
* [[Teach|Teach cluster]]&lt;br /&gt;
* [[FAQ | FAQ (frequently asked questions)]]&lt;br /&gt;
* [[Acknowledging SciNet]]&lt;br /&gt;
| valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== Tutorials, Manuals, etc. ==&lt;br /&gt;
* [https://education.scinet.utoronto.ca SciNet education material]&lt;br /&gt;
* [https://www.youtube.com/c/SciNetHPCattheUniversityofToronto SciNet's YouTube channel]&lt;br /&gt;
* [[Modules specific to Niagara|Software Modules specific to Niagara]] &lt;br /&gt;
* [[Modules for Mist]] &lt;br /&gt;
* [[Commercial software]]&lt;br /&gt;
* [[Burst Buffer]]&lt;br /&gt;
* [[SSH#SSH Keys|SSH keys]]&lt;br /&gt;
* [[SSH Tunneling]]&lt;br /&gt;
* [[Visualization]]&lt;br /&gt;
* [[Running Serial Jobs on Niagara]]&lt;br /&gt;
* [[Jupyter Hub]]&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Rzon</name></author>
	</entry>
	<entry>
		<id>https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7535</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7535"/>
		<updated>2026-02-17T21:03:44Z</updated>

		<summary type="html">&lt;p&gt;Rzon: /* System Status */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;__NOTOC__&lt;br /&gt;
{| style=&amp;quot;border-spacing:10px; width: 95%&amp;quot;&lt;br /&gt;
| style=&amp;quot;padding:1em; padding-top:.1em; border:2px solid #0645ad; background-color:#f6f6f6; border-radius:7px&amp;quot;|&lt;br /&gt;
&lt;br /&gt;
==System Status==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- Use &amp;quot;Up&amp;quot;, &amp;quot;Partial&amp;quot; or &amp;quot;Down&amp;quot;; these are templates. --&amp;gt;&lt;br /&gt;
{|style=&amp;quot;width:100%&amp;quot; &lt;br /&gt;
|{{Down3 | Trillium|https://docs.alliancecan.ca/wiki/Trillium_Quickstart}}&lt;br /&gt;
|{{Up3 | OnDemand|https://docs.alliancecan.ca/wiki/Trillium_Open_OnDemand_Quickstart}}&lt;br /&gt;
|{{Up | Globus |Globus}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Up | HPSS|HPSS}}&lt;br /&gt;
|{{Up | Balam|Balam}}&lt;br /&gt;
|{{Partial | S4H | S4H}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Up | Teach|Teach}}&lt;br /&gt;
|{{Up3 | File system|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Storage}}&lt;br /&gt;
|{{Up3 | External Network|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Logging_in}} &lt;br /&gt;
|}&lt;br /&gt;
'''Tue Feb 17, 2026, 8:40 am:''' Power outage at the data centre.  Cooling issues have developed as a result.  Major systems (Trillium, S4H) are expected to be down until sometime Thursday.&lt;br /&gt;
&lt;br /&gt;
'''Mon Feb 16, 2026, 8:40 pm:''' Electricity is unstable in the data centre area due to severe snowfall.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 1:40 pm:''' All services are operational again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 12:00 pm:''' The Trillium and Open OnDemand compute nodes are operational again. We are still working on bringing Balam, Neptune and S4H nodes up again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 10:00 am:''' There was a power glitch at the data centre overnight. The login nodes are accessible but the compute nodes are down.  &lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 16, 2026, 11:00 pm:''' HPSS is back online, and accessible via alliancecan#hpss Globus endpoint. &lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 15, 2026, 10:00 pm:''' HPSS will undergo maintenance on Friday morning, Jan/16/2025, , including alliancecan#hpss Globus endpoint &lt;br /&gt;
&lt;br /&gt;
'''Tue Jan 6, 2026, 10:15 am:''' OnDemand has been fixed and is working again.&lt;br /&gt;
&lt;br /&gt;
'''Mon Jan 5, 2026, 9:00 pm:''' The authentication mechanism of OnDemand is not working.&lt;br /&gt;
&lt;br /&gt;
'''Wed Dec 31, 2025, 12:40 pm:''' We believe the problem has now been resolved.  Please let us know if you still experience login problems or aborted jobs.&lt;br /&gt;
&lt;br /&gt;
'''Tue Dec 30, 2025, 2:10 pm:''' We are experiencing problems with authentication, resulting in failed logins, OOD errors, and aborted jobs (with &amp;quot;prolog error&amp;quot;).  Please bear with us, as we are very short-staffed during the holiday break.  We will post updates here.&lt;br /&gt;
&lt;br /&gt;
'''Tue Dec 3, 2025, 11:30 am:''' Open OnDemand is fully operational again.&lt;br /&gt;
&lt;br /&gt;
'''Sat Nov 29, 2025, 00:40 am:''' There has been a problem with the water chiller. Some systems are offline.&lt;br /&gt;
&lt;br /&gt;
'''Wed Nov 5, 2025, 12:55 pm:''' Balam is back online.&lt;br /&gt;
&lt;br /&gt;
'''Wed Nov 5, 2025, 10:00 am:''' Open OnDemand is back online.&lt;br /&gt;
&lt;br /&gt;
'''Tue Nov 4, 2025, 11:00 pm:''' Most of the work is done, data movers, Globus, and HPSS are back online. Remaining services will be worked on tomorrow.&lt;br /&gt;
&lt;br /&gt;
'''Tue Nov 4, 2025, 8:30 am:''' Scheduled network maintenance. Trillium cluster is *not* affected.&lt;br /&gt;
&lt;br /&gt;
'''Tue Oct 21, 2025, 17:30 am:''' Balam maintenance finished.&lt;br /&gt;
&lt;br /&gt;
'''Tue Oct 21, 2025, 7:00 am:''' Balam maintenance day.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 15, 2025, 3:55 pm:''' Trillium inbound connections through trillium.alliancecan.ca or trillium.scinet.utoronto.ca are working again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 15, 2025, 3:05 pm:''' Trillium is experiencing external network issues for both incoming traffic. Please try: ssh USERNAME@tri-login01.scinet.utoronto.ca in the meantime.&lt;br /&gt;
 &lt;br /&gt;
'''Thu Oct 06, 2025, 8:00 pm:''' HPSS is fully functional. You may submit archive jobs from trillium login nodes, datamovers and robots.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 03, 2025, 6:30 pm:''' HPSS is back online, and already accessible via alliancecan#hpss Globus endpoint. Directory tree now follows the other Alliance clusters. We're still working on job submission via Slurm&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 01, 2025, 0:00 am:''' Niagara compute nodes are now unavailable for regular users. The login nodes will remain available for a while to allow a few last data transfers, although transfers from the Niagara file systems to Trillium are best done on nia-dm1.scinet.utoronto.ca.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 01, 2025, 9:30 am:''' HPSS is down for scheduled maintenance, including alliancecan#hpss Globus endpoint&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--  When removing system status entries, please archive them to: --&amp;gt;&lt;br /&gt;
[[Previous messages]]&lt;br /&gt;
&lt;br /&gt;
{|style=&amp;quot;border-spacing: 10px;width: 100%&amp;quot;&lt;br /&gt;
|valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== QuickStart Guides ==&lt;br /&gt;
* [https://docs.alliancecan.ca/wiki/Trillium_Quickstart Trillium Quickstart]&lt;br /&gt;
* [[Niagara Quickstart]]&lt;br /&gt;
* [[HPSS | HPSS archival storage]]&lt;br /&gt;
* [[Teach|Teach cluster]]&lt;br /&gt;
* [[FAQ | FAQ (frequently asked questions)]]&lt;br /&gt;
* [[Acknowledging SciNet]]&lt;br /&gt;
| valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== Tutorials, Manuals, etc. ==&lt;br /&gt;
* [https://education.scinet.utoronto.ca SciNet education material]&lt;br /&gt;
* [https://www.youtube.com/c/SciNetHPCattheUniversityofToronto SciNet's YouTube channel]&lt;br /&gt;
* [[Modules specific to Niagara|Software Modules specific to Niagara]] &lt;br /&gt;
* [[Modules for Mist]] &lt;br /&gt;
* [[Commercial software]]&lt;br /&gt;
* [[Burst Buffer]]&lt;br /&gt;
* [[SSH#SSH Keys|SSH keys]]&lt;br /&gt;
* [[SSH Tunneling]]&lt;br /&gt;
* [[Visualization]]&lt;br /&gt;
* [[Running Serial Jobs on Niagara]]&lt;br /&gt;
* [[Jupyter Hub]]&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Rzon</name></author>
	</entry>
	<entry>
		<id>https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7529</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7529"/>
		<updated>2026-02-17T03:49:23Z</updated>

		<summary type="html">&lt;p&gt;Rzon: /* System Status */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;__NOTOC__&lt;br /&gt;
{| style=&amp;quot;border-spacing:10px; width: 95%&amp;quot;&lt;br /&gt;
| style=&amp;quot;padding:1em; padding-top:.1em; border:2px solid #0645ad; background-color:#f6f6f6; border-radius:7px&amp;quot;|&lt;br /&gt;
&lt;br /&gt;
==System Status==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- Use &amp;quot;Up&amp;quot;, &amp;quot;Partial&amp;quot; or &amp;quot;Down&amp;quot;; these are templates. --&amp;gt;&lt;br /&gt;
{|style=&amp;quot;width:100%&amp;quot; &lt;br /&gt;
|{{Partial3 | Trillium|https://docs.alliancecan.ca/wiki/Trillium_Quickstart}}&lt;br /&gt;
|{{Partial3 | OnDemand|https://docs.alliancecan.ca/wiki/Trillium_Open_OnDemand_Quickstart}}&lt;br /&gt;
|{{Up | Globus |Globus}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Up | HPSS|HPSS}}&lt;br /&gt;
|{{Partial | Balam|Balam}}&lt;br /&gt;
|{{Partial | S4H | S4H}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Partial | Teach|Teach}}&lt;br /&gt;
|{{Up3 | File system|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Storage}}&lt;br /&gt;
|{{Up3 | External Network|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Logging_in}} &lt;br /&gt;
|}&lt;br /&gt;
'''Thu Feb 16, 2026, 8:40 pm:''' Electricity is unstable in the data centre area due to severe snowfall.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 1:40 pm:''' All services are operational again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 12:00 pm:''' The Trillium and Open OnDemand compute nodes are operational again. We are still working on bringing Balam, Neptune and S4H nodes up again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 10:00 am:''' There was a power glitch at the data centre overnight. The login nodes are accessible but the compute nodes are down.  &lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 16, 2026, 11:00 pm:''' HPSS is back online, and accessible via alliancecan#hpss Globus endpoint. &lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 15, 2026, 10:00 pm:''' HPSS will undergo maintenance on Friday morning, Jan/16/2025, , including alliancecan#hpss Globus endpoint &lt;br /&gt;
&lt;br /&gt;
'''Tue Jan 6, 2026, 10:15 am:''' OnDemand has been fixed and is working again.&lt;br /&gt;
&lt;br /&gt;
'''Mon Jan 5, 2026, 9:00 pm:''' The authentication mechanism of OnDemand is not working.&lt;br /&gt;
&lt;br /&gt;
'''Wed Dec 31, 2025, 12:40 pm:''' We believe the problem has now been resolved.  Please let us know if you still experience login problems or aborted jobs.&lt;br /&gt;
&lt;br /&gt;
'''Tue Dec 30, 2025, 2:10 pm:''' We are experiencing problems with authentication, resulting in failed logins, OOD errors, and aborted jobs (with &amp;quot;prolog error&amp;quot;).  Please bear with us, as we are very short-staffed during the holiday break.  We will post updates here.&lt;br /&gt;
&lt;br /&gt;
'''Tue Dec 3, 2025, 11:30 am:''' Open OnDemand is fully operational again.&lt;br /&gt;
&lt;br /&gt;
'''Sat Nov 29, 2025, 00:40 am:''' There has been a problem with the water chiller. Some systems are offline.&lt;br /&gt;
&lt;br /&gt;
'''Wed Nov 5, 2025, 12:55 pm:''' Balam is back online.&lt;br /&gt;
&lt;br /&gt;
'''Wed Nov 5, 2025, 10:00 am:''' Open OnDemand is back online.&lt;br /&gt;
&lt;br /&gt;
'''Tue Nov 4, 2025, 11:00 pm:''' Most of the work is done, data movers, Globus, and HPSS are back online. Remaining services will be worked on tomorrow.&lt;br /&gt;
&lt;br /&gt;
'''Tue Nov 4, 2025, 8:30 am:''' Scheduled network maintenance. Trillium cluster is *not* affected.&lt;br /&gt;
&lt;br /&gt;
'''Tue Oct 21, 2025, 17:30 am:''' Balam maintenance finished.&lt;br /&gt;
&lt;br /&gt;
'''Tue Oct 21, 2025, 7:00 am:''' Balam maintenance day.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 15, 2025, 3:55 pm:''' Trillium inbound connections through trillium.alliancecan.ca or trillium.scinet.utoronto.ca are working again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 15, 2025, 3:05 pm:''' Trillium is experiencing external network issues for both incoming traffic. Please try: ssh USERNAME@tri-login01.scinet.utoronto.ca in the meantime.&lt;br /&gt;
 &lt;br /&gt;
'''Thu Oct 06, 2025, 8:00 pm:''' HPSS is fully functional. You may submit archive jobs from trillium login nodes, datamovers and robots.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 03, 2025, 6:30 pm:''' HPSS is back online, and already accessible via alliancecan#hpss Globus endpoint. Directory tree now follows the other Alliance clusters. We're still working on job submission via Slurm&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 01, 2025, 0:00 am:''' Niagara compute nodes are now unavailable for regular users. The login nodes will remain available for a while to allow a few last data transfers, although transfers from the Niagara file systems to Trillium are best done on nia-dm1.scinet.utoronto.ca.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 01, 2025, 9:30 am:''' HPSS is down for scheduled maintenance, including alliancecan#hpss Globus endpoint&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--  When removing system status entries, please archive them to: --&amp;gt;&lt;br /&gt;
[[Previous messages]]&lt;br /&gt;
&lt;br /&gt;
{|style=&amp;quot;border-spacing: 10px;width: 100%&amp;quot;&lt;br /&gt;
|valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== QuickStart Guides ==&lt;br /&gt;
* [https://docs.alliancecan.ca/wiki/Trillium_Quickstart Trillium Quickstart]&lt;br /&gt;
* [[Niagara Quickstart]]&lt;br /&gt;
* [[HPSS | HPSS archival storage]]&lt;br /&gt;
* [[Teach|Teach cluster]]&lt;br /&gt;
* [[FAQ | FAQ (frequently asked questions)]]&lt;br /&gt;
* [[Acknowledging SciNet]]&lt;br /&gt;
| valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== Tutorials, Manuals, etc. ==&lt;br /&gt;
* [https://education.scinet.utoronto.ca SciNet education material]&lt;br /&gt;
* [https://www.youtube.com/c/SciNetHPCattheUniversityofToronto SciNet's YouTube channel]&lt;br /&gt;
* [[Modules specific to Niagara|Software Modules specific to Niagara]] &lt;br /&gt;
* [[Modules for Mist]] &lt;br /&gt;
* [[Commercial software]]&lt;br /&gt;
* [[Burst Buffer]]&lt;br /&gt;
* [[SSH#SSH Keys|SSH keys]]&lt;br /&gt;
* [[SSH Tunneling]]&lt;br /&gt;
* [[Visualization]]&lt;br /&gt;
* [[Running Serial Jobs on Niagara]]&lt;br /&gt;
* [[Jupyter Hub]]&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Rzon</name></author>
	</entry>
	<entry>
		<id>https://docs.scinet.utoronto.ca/index.php?title=Teach&amp;diff=7520</id>
		<title>Teach</title>
		<link rel="alternate" type="text/html" href="https://docs.scinet.utoronto.ca/index.php?title=Teach&amp;diff=7520"/>
		<updated>2026-02-13T17:25:35Z</updated>

		<summary type="html">&lt;p&gt;Rzon: /* Teaching Cluster */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Infobox Computer&lt;br /&gt;
|image=[[Image:Ibm_idataplex_dx360_m4.jpg|center|300px|thumb]] &lt;br /&gt;
|name=Teach Cluster &lt;br /&gt;
|installed=(orig Mar 2020), Feb 2025&lt;br /&gt;
|operatingsystem= Linux (Rocky 9.6)&lt;br /&gt;
|loginnode=teach-login01&lt;br /&gt;
|nnodes=64 &lt;br /&gt;
|rampernode=188 GiB / 202 GB &lt;br /&gt;
|corespernode=40 &lt;br /&gt;
|interconnect=Infiniband (EDR)&lt;br /&gt;
|vendorcompilers=gcc,intel&lt;br /&gt;
|queuetype=slurm&lt;br /&gt;
}}&lt;br /&gt;
&lt;br /&gt;
== Teaching Cluster ==&lt;br /&gt;
&lt;br /&gt;
The Teach cluster is a 2560-core cluster provided for teaching purposes.  It has currently no GPU capability.&lt;br /&gt;
It is configured similarly to [https://scinethpc.ca/trillium/  Trillium], however it uses hardware repurposed from its predecessor, [https://scinethpc.ca/niagara Niagara].  This system should not be used for production work as the queuing policies are designed to provide fast job turnover and limit the amount of resources one person can use at a time.     Questions about its use or problems should be sent to '''support@scinet.utoronto.ca'''.&lt;br /&gt;
&lt;br /&gt;
== Specifications==&lt;br /&gt;
&lt;br /&gt;
This cluster currently consists of 64 repurposed x86_64 nodes each with 40 cores (from two 20 core Intel CascadeLake CPUs) running at 2.5GHz with 188GB of RAM per node. &lt;br /&gt;
The nodes are interconnected with 1:1 non-blocking EDR Infiniband for MPI communications, and disk I/O to a separate view of the VAST file system.  In total, this cluster contains 2560 cores.&lt;br /&gt;
&lt;br /&gt;
== Login/Devel Node ==&lt;br /&gt;
&lt;br /&gt;
Teach runs Rocky Linux 9.  You will need to be somewhat familiar with Linux systems to work on Teach.  If you are not, it will be worth your time to review our [https://education.scinet.utoronto.ca/tag/index.php?tag=SCMP101 Introduction to Linux Shell] class.&lt;br /&gt;
&lt;br /&gt;
As with all SciNet and {{Alliance}} systems, access to Teach is done via [[SSH]] (secure shell) only.  Open a terminal window (e.g. using [https://docs.alliancecan.ca/wiki/Connecting_with_PuTTY PuTTY] or  [https://docs.alliancecan.ca/wiki/Connecting_with_MobaXTerm MobaXTerm] on Windows), and type&lt;br /&gt;
 ssh -Y USERNAME@teach.scinet.utoronto.ca&lt;br /&gt;
This will bring directly to the command line of '''&amp;lt;tt&amp;gt;teach-login01&amp;lt;/tt&amp;gt;''' or '''&amp;lt;tt&amp;gt;teach-login02&amp;lt;/tt&amp;gt;''', which are the gateway/devel nodes for this cluster.  &lt;br /&gt;
On these nodes, you can compile, do short tests, and submit your jobs to the queue.&lt;br /&gt;
&lt;br /&gt;
The first time you login to Teach cluster, please make sure to check if the login node ssh key fingerprint&lt;br /&gt;
matches. [[Teach_fingerprints | See here how]].&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The login nodes are shared between students of a          &lt;br /&gt;
number of different courses. Use this node to develop and compile                  &lt;br /&gt;
code, to run short tests, and to submit computations to the scheduler (see below).&lt;br /&gt;
&lt;br /&gt;
Note that access to the Teach cluster is restricted to temporary accounts that start with the prefix '''lcl_uot''' + the course code + '''s''', and a number.  Passwords for these accounts can be changed on the [https://portal.scinet.utoronto.ca/portaluserlogin SciNet user portal]. On the same site, users can upload a public ssh key if they want to connect using ssh keys.&lt;br /&gt;
&lt;br /&gt;
== Software Modules ==&lt;br /&gt;
&lt;br /&gt;
Other than essentials, all installed software is made available [[Using_modules | using module commands]]. These modules set environment variables (PATH, etc.), allowing multiple, conflicting versions of a given package to be available.  A detailed explanation of the module system can be [[Using_modules | found on the modules page]].  &lt;br /&gt;
&lt;br /&gt;
The Teach cluster makes the same [https://docs.alliancecan.ca/wiki/Available_software modules available] as on the [https://docs.alliancecan.ca General Purpose clusters of the Digital Research Alliance of Canada], with one caveat.  On Teach, by default, only the &amp;quot;gentoo&amp;quot; module is loaded, which provides basic OS-level functionality.  &lt;br /&gt;
&lt;br /&gt;
Common module subcommands are:&lt;br /&gt;
&lt;br /&gt;
* &amp;lt;code&amp;gt;module load &amp;lt;module-name&amp;gt;&amp;lt;/code&amp;gt;: load the default version of a particular software.&lt;br /&gt;
* &amp;lt;code&amp;gt;module load &amp;lt;module-name&amp;gt;/&amp;lt;module-version&amp;gt;&amp;lt;/code&amp;gt;: load a specific version of a particular software.&lt;br /&gt;
* &amp;lt;code&amp;gt;module purge&amp;lt;/code&amp;gt;: unload all currently loaded modules.&lt;br /&gt;
* &amp;lt;code&amp;gt;module spider&amp;lt;/code&amp;gt; (or &amp;lt;code&amp;gt;module spider &amp;lt;module-name&amp;gt;&amp;lt;/code&amp;gt;): list available software packages.&lt;br /&gt;
* &amp;lt;code&amp;gt;module avail&amp;lt;/code&amp;gt;: list loadable software packages.&lt;br /&gt;
* &amp;lt;code&amp;gt;module list&amp;lt;/code&amp;gt;: list loaded modules.&lt;br /&gt;
&lt;br /&gt;
For example, to make the GNU compilers (gcc, g++ and gfortran) available, you should type&lt;br /&gt;
&lt;br /&gt;
 module load gcc&lt;br /&gt;
&lt;br /&gt;
while the Intel compilers (icc, icpc and ifort) can be loaded by&lt;br /&gt;
&lt;br /&gt;
 module load intel&lt;br /&gt;
&lt;br /&gt;
To get the default modules that are loaded on the General Purpose clusters, you can load the &amp;quot;StdEnv&amp;quot; module. &lt;br /&gt;
&lt;br /&gt;
Along with modifying common environment variables, such as the PATH, these modules also create an '''EBROOT&amp;lt;MODULENAME&amp;gt;''' environment variable, which can be used to access commonly needed software directories, such as /include and /lib.&lt;br /&gt;
&lt;br /&gt;
There are handy abbreviations for the module commands. &amp;lt;code&amp;gt;ml&amp;lt;/code&amp;gt; is the same as &amp;lt;code&amp;gt;module list&amp;lt;/code&amp;gt;, and &amp;lt;code&amp;gt;ml &amp;lt;module-name&amp;gt;&amp;lt;/code&amp;gt; is the same as &amp;lt;code&amp;gt;module load &amp;lt;module-name&amp;gt;&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
A list of available software modules can be on [https://docs.alliancecan.ca/wiki/Available_software found on this page].&lt;br /&gt;
&lt;br /&gt;
There are a few addition modules available as well, and more can be made available upon demand of the course instructors.  Currently, the only additional modules are:&lt;br /&gt;
&lt;br /&gt;
 catch2/3.3.1      - A C++ test framework for unit-tests, TDD and BDD using C++14 and later.&lt;br /&gt;
 misopy/0.5.2      - A probabilistic framework to analysize RNA-Seq data.&lt;br /&gt;
 palemoon/33.6.0.1 - A web browser&lt;br /&gt;
&lt;br /&gt;
== Interactive jobs ==&lt;br /&gt;
&lt;br /&gt;
For a interactive sessions on a compute node of the teach cluster that give access to non-shared resources, use the 'debugjob' command. &lt;br /&gt;
 teach01:~$ debugjob -n C&lt;br /&gt;
where C is the number of cores. An interactive session defaults to four hours when using at most one node (C&amp;lt;=40), and becomes 60 minutes when using four nodes (i.e., 120&amp;lt;C&amp;lt;=160), which is the maximum number of nodes allowed for an interactive session by debugjob.&lt;br /&gt;
                                                             &lt;br /&gt;
For a short interactive sessions on a dedicated compute node of the teach cluster, use the 'debugjob' command as follows: &lt;br /&gt;
 teach01:~$ debugjob N&lt;br /&gt;
where N is the number of nodes.  On the Teach cluster, this is equivalent to &amp;lt;tt&amp;gt;debugjob -n 40*N &amp;lt;/tt&amp;gt;. The positive integer number &amp;lt;tt&amp;gt;N&amp;lt;/tt&amp;gt; can at most be 4.&lt;br /&gt;
&lt;br /&gt;
If no arguments are given to &amp;lt;tt&amp;gt;debugjob&amp;lt;/tt&amp;gt;, it allocates a single core on a Teach compute node.&lt;br /&gt;
&lt;br /&gt;
There are limits on the resources you can get with a debugjob, and how long you can get them.  No debugjob can run longer than four hours or use more than 160 cores, and each user can only run one at a time.  For longer computations, jobs must be submitted to the scheduler.&lt;br /&gt;
&lt;br /&gt;
== Submit a Job ==&lt;br /&gt;
&lt;br /&gt;
Teach uses SLURM as its job scheduler.  More-advanced details of how to interact with the scheduler can be found on the [[Slurm | Slurm page]].&lt;br /&gt;
&lt;br /&gt;
You submit jobs from a login node by passing a script to the sbatch command:&lt;br /&gt;
&lt;br /&gt;
 teach-login01:~$ sbatch jobscript.sh&lt;br /&gt;
&lt;br /&gt;
This puts the job in the queue. It will run on the compute nodes in due course.&lt;br /&gt;
&lt;br /&gt;
Note:&lt;br /&gt;
* Make sure to adjust accordingly the flags --ntasks-per-node or --ntasks together with --nodes for the examples found at [[Slurm | Slurm page]]. &lt;br /&gt;
* The current slurm configuration of the teach cluster allocates compute resources by core as opposed to by node. That means your tasks might land on nodes that have other jobs running, i.e. they might share the node. If you want to avoid that, make sure to add the following directive in your submitting script: #SBATCH --exclusive. This forces your job to use the compute nodes exclusively.&lt;br /&gt;
* The maximum wall time is currently set to 4 hours.&lt;br /&gt;
* There are 2 queues available: Compute queue and debug queue. Their usage limits are listed on the table below.&lt;br /&gt;
&lt;br /&gt;
== Limits ==&lt;br /&gt;
There are limits to the size and duration of your jobs, the number of jobs you can run and the number of jobs you can have queued. It also matters in which 'partition' the jobs runs. 'Partitions' are SLURM-speak for use cases. You specify the partition with the -p parameter to sbatch or salloc, but if you do not specify one, your job will run in the compute partition, which is the most common case.&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!Usage&lt;br /&gt;
!Partition&lt;br /&gt;
!Running jobs&lt;br /&gt;
!Submitted jobs (incl. running)&lt;br /&gt;
!Min. size of jobs&lt;br /&gt;
!Max. size of jobs&lt;br /&gt;
!Min. walltime&lt;br /&gt;
!Max. walltime &lt;br /&gt;
|-&lt;br /&gt;
|Interactive testing or troubleshooting || debug || 1 || 1 || 1 core || 4 nodes (160 cores)|| N/A || 4 hours&lt;br /&gt;
|-&lt;br /&gt;
|Compute jobs ||compute || 1 || 12 || 1 core || 4 nodes (160 cores)|| 15 minutes || 4 hours&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
Within these limits, jobs may still have to wait in the queue.  Although there are no allocations on the teach cluster, the waiting time still depends on many factors, such as the number of nodes and the wall time, how many other jobs are waiting in the queue, and whether a job can fill an otherwise unused spot in the schedule.&lt;br /&gt;
&lt;br /&gt;
== Main changes from Teach's predecessor ==&lt;br /&gt;
&lt;br /&gt;
This section is intended for instructors that may have use SciNet's previous Teach cluster.&lt;br /&gt;
&lt;br /&gt;
This Teach cluster is setup differently from its predecessor.  &lt;br /&gt;
&lt;br /&gt;
Although the cluster is once again called ''Teach'' and you connect with to teach.scinet.utoronto.ca as before, the system is setup differently from the previous Teach cluster in the following ways:&lt;br /&gt;
&lt;br /&gt;
* There are now 2 dedicated login nodes, teach-login01 and teach-login02.  &lt;br /&gt;
* The ssh fingerprints for these login nodes can be found on [[Teach_fingerprints]].&lt;br /&gt;
* The compute nodes have 40 cores.  As before, you can request jobs by number of cores.&lt;br /&gt;
* Only temporary lcl_uot.... accounts can log in.&lt;br /&gt;
* Only the home directories of those account are mounted. &lt;br /&gt;
* In particular, the file systems from the other SciNet compute clusters (Niagara, Mist, Trillium,... ) are not and will not be mounted.  You'll need to copy over any files that you need to use on the Teach cluster.&lt;br /&gt;
* There is no $SCRATCH.  You can do all your work on $HOME, which is writable from compute nodes&lt;br /&gt;
* The software stack is the one supplied by the Alliance.  There is no need to load 'CCEnv' to get them.&lt;br /&gt;
* But as before, if you're missing a module, we can still install it for you.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- Do this later&lt;br /&gt;
== Running Jupyter on a Teach Compute Node ==  &lt;br /&gt;
&lt;br /&gt;
1. To be able to run Jupyter on a compute node, you must first (a) install it inside a virtual environment, (b) enable a way for jupyter to seemingly write to a specific directory on $HOME, and (c) create a little helper script called &amp;lt;tt&amp;gt;notebook.sh&amp;lt;/tt&amp;gt; that will be used to start the jupyter server in step 2.  These are the commands that you should use for the installation (which you should do only once, on the Teach login node):&lt;br /&gt;
&lt;br /&gt;
(a) Create virtual env&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
$ module load python/3.9.10&lt;br /&gt;
$ virtualenv --system-site-packages $HOME/.virtualenvs/jupteach&lt;br /&gt;
$ source $HOME/.virtualenvs/jupteach/bin/activate&lt;br /&gt;
$ pip install jupyter jupyterlab&lt;br /&gt;
$ deactivate&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
You can choose another directory than &amp;lt;tt&amp;gt;$HOME/.virtualenvs/jupteach&amp;lt;/tt&amp;gt; for where to create the virtual environment, but you need to be consistent and use the same directory everywhere below.&lt;br /&gt;
&lt;br /&gt;
(b) Make a writable 'runtime' directory for Jupyter. &lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
$ mkdir -p $HOME/.local/share/jupyter/runtime &lt;br /&gt;
$ mv -f $HOME/.local/share/jupyter/runtime $SCRATCH/jupyter_runtime || mkdir $SCRATCH/jupyter_runtime&lt;br /&gt;
$ ln -sT $SCRATCH/jupyter_runtime $HOME/.local/share/jupyter/runtime&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
(c) Create a launch script to use on the compute nodes:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
$ cat &amp;gt; $HOME/.virtualenvs/jupteach/bin/notebook.sh &amp;lt;&amp;lt;EOF&lt;br /&gt;
#!/bin/bash&lt;br /&gt;
source \$HOME/.virtualenvs/jupteach/bin/activate&lt;br /&gt;
export XDG_DATA_HOME=\$SCRATCH/.share&lt;br /&gt;
export XDG_CACHE_HOME=\$SCRATCH/.cache&lt;br /&gt;
export XDG_CONFIG_HOME=\$SCRATCH/.config&lt;br /&gt;
export XDG_RUNTIME_DIR=\$SCRATCH/.runtime&lt;br /&gt;
export JUPYTER_CONFIG_DIR=\$SCRATCH/.config/.jupyter&lt;br /&gt;
jupyter \${1:-notebook} --ip \$(hostname -f) --no-browser --notebook-dir=\$PWD&lt;br /&gt;
EOF&lt;br /&gt;
$ chmod +x  $HOME/.virtualenvs/jupteach/bin/notebook.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2. To run the jupyter server on a compute node, start an interactive session with the &amp;lt;tt&amp;gt;debugjob&amp;lt;/tt&amp;gt; command and then launch the jupyter server:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
$ debugjob -n 16                    # use less if you need less cores.&lt;br /&gt;
$ cd $SCRATCH                       # $HOME is read-only, so move to $SCRATCH&lt;br /&gt;
$ $HOME/.virtualenvs/jupteach/bin/notebook.sh  # add the argument &amp;quot;lab&amp;quot; to start with the jupyter lab&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
Make sure you note down (a) the name of the compute node that you got allocated (they start with &amp;quot;&amp;lt;tt&amp;gt;teach&amp;lt;/tt&amp;gt;&amp;quot; followed by a 2-digit number), and (b) the number following the compute nodes after the colon (usually this is 8888, but it can be another, higher number); this is the PORT  and (c) the last URL that the notebook.sh tells you to use to connect.&lt;br /&gt;
&lt;br /&gt;
4. To connect to this jupyter server running on a teach compute node, which is not accessible from the internet, in a different terminal on you own computer, you must reconnect to the Teach cluster with a port-forwarding&lt;br /&gt;
tunnel to the compute node on which jupyter is running:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
$ ssh -LPORT:teachXX:PORT -o ControlMaster=no USERNAME@teach.scinet.utoronto.ca  -N&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;tt&amp;gt;teachXX&amp;lt;/tt&amp;gt; is to be replaced by the name of the compute node (point (a) above), PORT is to be replaced by the port number that notebook.sh showed, and &amp;lt;tt&amp;gt;USERNAME&amp;lt;/t&amp;gt; should be your teach account username. This command should just &amp;quot;hang&amp;quot; there, it only serves to forward port number PORT (usually 8888) to port PORT (usually 8888) on the compute node.  &lt;br /&gt;
&lt;br /&gt;
Finally, point your browser to the URL that the &amp;lt;tt&amp;gt;notebook.sh&amp;lt;/tt&amp;gt; command printed out (point (b) above), i.e., the one with 127.0.0.1 in it.&lt;br /&gt;
--&amp;gt;&lt;/div&gt;</summary>
		<author><name>Rzon</name></author>
	</entry>
	<entry>
		<id>https://docs.scinet.utoronto.ca/index.php?title=Teach&amp;diff=7517</id>
		<title>Teach</title>
		<link rel="alternate" type="text/html" href="https://docs.scinet.utoronto.ca/index.php?title=Teach&amp;diff=7517"/>
		<updated>2026-02-13T17:25:13Z</updated>

		<summary type="html">&lt;p&gt;Rzon: /* Teaching Cluster */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Infobox Computer&lt;br /&gt;
|image=[[Image:Ibm_idataplex_dx360_m4.jpg|center|300px|thumb]] &lt;br /&gt;
|name=Teach Cluster &lt;br /&gt;
|installed=(orig Mar 2020), Feb 2025&lt;br /&gt;
|operatingsystem= Linux (Rocky 9.6)&lt;br /&gt;
|loginnode=teach-login01&lt;br /&gt;
|nnodes=64 &lt;br /&gt;
|rampernode=188 GiB / 202 GB &lt;br /&gt;
|corespernode=40 &lt;br /&gt;
|interconnect=Infiniband (EDR)&lt;br /&gt;
|vendorcompilers=gcc,intel&lt;br /&gt;
|queuetype=slurm&lt;br /&gt;
}}&lt;br /&gt;
&lt;br /&gt;
== Teaching Cluster ==&lt;br /&gt;
&lt;br /&gt;
The Teach cluster is a 2560-core cluster provided for teaching purposes.  It has currently no GPU capability.&lt;br /&gt;
It is configured similarly to [https://scinethpc.ca/trillium/  Trillium], however it uses hardware repurposed from its predecessor, [[https://scinethpc.ca/niagara]].  This system should not be used for production work as the queuing policies are designed to provide fast job turnover and limit the amount of resources one person can use at a time.     Questions about its use or problems should be sent to '''support@scinet.utoronto.ca'''.&lt;br /&gt;
&lt;br /&gt;
== Specifications==&lt;br /&gt;
&lt;br /&gt;
This cluster currently consists of 64 repurposed x86_64 nodes each with 40 cores (from two 20 core Intel CascadeLake CPUs) running at 2.5GHz with 188GB of RAM per node. &lt;br /&gt;
The nodes are interconnected with 1:1 non-blocking EDR Infiniband for MPI communications, and disk I/O to a separate view of the VAST file system.  In total, this cluster contains 2560 cores.&lt;br /&gt;
&lt;br /&gt;
== Login/Devel Node ==&lt;br /&gt;
&lt;br /&gt;
Teach runs Rocky Linux 9.  You will need to be somewhat familiar with Linux systems to work on Teach.  If you are not, it will be worth your time to review our [https://education.scinet.utoronto.ca/tag/index.php?tag=SCMP101 Introduction to Linux Shell] class.&lt;br /&gt;
&lt;br /&gt;
As with all SciNet and {{Alliance}} systems, access to Teach is done via [[SSH]] (secure shell) only.  Open a terminal window (e.g. using [https://docs.alliancecan.ca/wiki/Connecting_with_PuTTY PuTTY] or  [https://docs.alliancecan.ca/wiki/Connecting_with_MobaXTerm MobaXTerm] on Windows), and type&lt;br /&gt;
 ssh -Y USERNAME@teach.scinet.utoronto.ca&lt;br /&gt;
This will bring directly to the command line of '''&amp;lt;tt&amp;gt;teach-login01&amp;lt;/tt&amp;gt;''' or '''&amp;lt;tt&amp;gt;teach-login02&amp;lt;/tt&amp;gt;''', which are the gateway/devel nodes for this cluster.  &lt;br /&gt;
On these nodes, you can compile, do short tests, and submit your jobs to the queue.&lt;br /&gt;
&lt;br /&gt;
The first time you login to Teach cluster, please make sure to check if the login node ssh key fingerprint&lt;br /&gt;
matches. [[Teach_fingerprints | See here how]].&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The login nodes are shared between students of a          &lt;br /&gt;
number of different courses. Use this node to develop and compile                  &lt;br /&gt;
code, to run short tests, and to submit computations to the scheduler (see below).&lt;br /&gt;
&lt;br /&gt;
Note that access to the Teach cluster is restricted to temporary accounts that start with the prefix '''lcl_uot''' + the course code + '''s''', and a number.  Passwords for these accounts can be changed on the [https://portal.scinet.utoronto.ca/portaluserlogin SciNet user portal]. On the same site, users can upload a public ssh key if they want to connect using ssh keys.&lt;br /&gt;
&lt;br /&gt;
== Software Modules ==&lt;br /&gt;
&lt;br /&gt;
Other than essentials, all installed software is made available [[Using_modules | using module commands]]. These modules set environment variables (PATH, etc.), allowing multiple, conflicting versions of a given package to be available.  A detailed explanation of the module system can be [[Using_modules | found on the modules page]].  &lt;br /&gt;
&lt;br /&gt;
The Teach cluster makes the same [https://docs.alliancecan.ca/wiki/Available_software modules available] as on the [https://docs.alliancecan.ca General Purpose clusters of the Digital Research Alliance of Canada], with one caveat.  On Teach, by default, only the &amp;quot;gentoo&amp;quot; module is loaded, which provides basic OS-level functionality.  &lt;br /&gt;
&lt;br /&gt;
Common module subcommands are:&lt;br /&gt;
&lt;br /&gt;
* &amp;lt;code&amp;gt;module load &amp;lt;module-name&amp;gt;&amp;lt;/code&amp;gt;: load the default version of a particular software.&lt;br /&gt;
* &amp;lt;code&amp;gt;module load &amp;lt;module-name&amp;gt;/&amp;lt;module-version&amp;gt;&amp;lt;/code&amp;gt;: load a specific version of a particular software.&lt;br /&gt;
* &amp;lt;code&amp;gt;module purge&amp;lt;/code&amp;gt;: unload all currently loaded modules.&lt;br /&gt;
* &amp;lt;code&amp;gt;module spider&amp;lt;/code&amp;gt; (or &amp;lt;code&amp;gt;module spider &amp;lt;module-name&amp;gt;&amp;lt;/code&amp;gt;): list available software packages.&lt;br /&gt;
* &amp;lt;code&amp;gt;module avail&amp;lt;/code&amp;gt;: list loadable software packages.&lt;br /&gt;
* &amp;lt;code&amp;gt;module list&amp;lt;/code&amp;gt;: list loaded modules.&lt;br /&gt;
&lt;br /&gt;
For example, to make the GNU compilers (gcc, g++ and gfortran) available, you should type&lt;br /&gt;
&lt;br /&gt;
 module load gcc&lt;br /&gt;
&lt;br /&gt;
while the Intel compilers (icc, icpc and ifort) can be loaded by&lt;br /&gt;
&lt;br /&gt;
 module load intel&lt;br /&gt;
&lt;br /&gt;
To get the default modules that are loaded on the General Purpose clusters, you can load the &amp;quot;StdEnv&amp;quot; module. &lt;br /&gt;
&lt;br /&gt;
Along with modifying common environment variables, such as the PATH, these modules also create an '''EBROOT&amp;lt;MODULENAME&amp;gt;''' environment variable, which can be used to access commonly needed software directories, such as /include and /lib.&lt;br /&gt;
&lt;br /&gt;
There are handy abbreviations for the module commands. &amp;lt;code&amp;gt;ml&amp;lt;/code&amp;gt; is the same as &amp;lt;code&amp;gt;module list&amp;lt;/code&amp;gt;, and &amp;lt;code&amp;gt;ml &amp;lt;module-name&amp;gt;&amp;lt;/code&amp;gt; is the same as &amp;lt;code&amp;gt;module load &amp;lt;module-name&amp;gt;&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
A list of available software modules can be on [https://docs.alliancecan.ca/wiki/Available_software found on this page].&lt;br /&gt;
&lt;br /&gt;
There are a few addition modules available as well, and more can be made available upon demand of the course instructors.  Currently, the only additional modules are:&lt;br /&gt;
&lt;br /&gt;
 catch2/3.3.1      - A C++ test framework for unit-tests, TDD and BDD using C++14 and later.&lt;br /&gt;
 misopy/0.5.2      - A probabilistic framework to analysize RNA-Seq data.&lt;br /&gt;
 palemoon/33.6.0.1 - A web browser&lt;br /&gt;
&lt;br /&gt;
== Interactive jobs ==&lt;br /&gt;
&lt;br /&gt;
For a interactive sessions on a compute node of the teach cluster that give access to non-shared resources, use the 'debugjob' command. &lt;br /&gt;
 teach01:~$ debugjob -n C&lt;br /&gt;
where C is the number of cores. An interactive session defaults to four hours when using at most one node (C&amp;lt;=40), and becomes 60 minutes when using four nodes (i.e., 120&amp;lt;C&amp;lt;=160), which is the maximum number of nodes allowed for an interactive session by debugjob.&lt;br /&gt;
                                                             &lt;br /&gt;
For a short interactive sessions on a dedicated compute node of the teach cluster, use the 'debugjob' command as follows: &lt;br /&gt;
 teach01:~$ debugjob N&lt;br /&gt;
where N is the number of nodes.  On the Teach cluster, this is equivalent to &amp;lt;tt&amp;gt;debugjob -n 40*N &amp;lt;/tt&amp;gt;. The positive integer number &amp;lt;tt&amp;gt;N&amp;lt;/tt&amp;gt; can at most be 4.&lt;br /&gt;
&lt;br /&gt;
If no arguments are given to &amp;lt;tt&amp;gt;debugjob&amp;lt;/tt&amp;gt;, it allocates a single core on a Teach compute node.&lt;br /&gt;
&lt;br /&gt;
There are limits on the resources you can get with a debugjob, and how long you can get them.  No debugjob can run longer than four hours or use more than 160 cores, and each user can only run one at a time.  For longer computations, jobs must be submitted to the scheduler.&lt;br /&gt;
&lt;br /&gt;
== Submit a Job ==&lt;br /&gt;
&lt;br /&gt;
Teach uses SLURM as its job scheduler.  More-advanced details of how to interact with the scheduler can be found on the [[Slurm | Slurm page]].&lt;br /&gt;
&lt;br /&gt;
You submit jobs from a login node by passing a script to the sbatch command:&lt;br /&gt;
&lt;br /&gt;
 teach-login01:~$ sbatch jobscript.sh&lt;br /&gt;
&lt;br /&gt;
This puts the job in the queue. It will run on the compute nodes in due course.&lt;br /&gt;
&lt;br /&gt;
Note:&lt;br /&gt;
* Make sure to adjust accordingly the flags --ntasks-per-node or --ntasks together with --nodes for the examples found at [[Slurm | Slurm page]]. &lt;br /&gt;
* The current slurm configuration of the teach cluster allocates compute resources by core as opposed to by node. That means your tasks might land on nodes that have other jobs running, i.e. they might share the node. If you want to avoid that, make sure to add the following directive in your submitting script: #SBATCH --exclusive. This forces your job to use the compute nodes exclusively.&lt;br /&gt;
* The maximum wall time is currently set to 4 hours.&lt;br /&gt;
* There are 2 queues available: Compute queue and debug queue. Their usage limits are listed on the table below.&lt;br /&gt;
&lt;br /&gt;
== Limits ==&lt;br /&gt;
There are limits to the size and duration of your jobs, the number of jobs you can run and the number of jobs you can have queued. It also matters in which 'partition' the jobs runs. 'Partitions' are SLURM-speak for use cases. You specify the partition with the -p parameter to sbatch or salloc, but if you do not specify one, your job will run in the compute partition, which is the most common case.&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!Usage&lt;br /&gt;
!Partition&lt;br /&gt;
!Running jobs&lt;br /&gt;
!Submitted jobs (incl. running)&lt;br /&gt;
!Min. size of jobs&lt;br /&gt;
!Max. size of jobs&lt;br /&gt;
!Min. walltime&lt;br /&gt;
!Max. walltime &lt;br /&gt;
|-&lt;br /&gt;
|Interactive testing or troubleshooting || debug || 1 || 1 || 1 core || 4 nodes (160 cores)|| N/A || 4 hours&lt;br /&gt;
|-&lt;br /&gt;
|Compute jobs ||compute || 1 || 12 || 1 core || 4 nodes (160 cores)|| 15 minutes || 4 hours&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
Within these limits, jobs may still have to wait in the queue.  Although there are no allocations on the teach cluster, the waiting time still depends on many factors, such as the number of nodes and the wall time, how many other jobs are waiting in the queue, and whether a job can fill an otherwise unused spot in the schedule.&lt;br /&gt;
&lt;br /&gt;
== Main changes from Teach's predecessor ==&lt;br /&gt;
&lt;br /&gt;
This section is intended for instructors that may have use SciNet's previous Teach cluster.&lt;br /&gt;
&lt;br /&gt;
This Teach cluster is setup differently from its predecessor.  &lt;br /&gt;
&lt;br /&gt;
Although the cluster is once again called ''Teach'' and you connect with to teach.scinet.utoronto.ca as before, the system is setup differently from the previous Teach cluster in the following ways:&lt;br /&gt;
&lt;br /&gt;
* There are now 2 dedicated login nodes, teach-login01 and teach-login02.  &lt;br /&gt;
* The ssh fingerprints for these login nodes can be found on [[Teach_fingerprints]].&lt;br /&gt;
* The compute nodes have 40 cores.  As before, you can request jobs by number of cores.&lt;br /&gt;
* Only temporary lcl_uot.... accounts can log in.&lt;br /&gt;
* Only the home directories of those account are mounted. &lt;br /&gt;
* In particular, the file systems from the other SciNet compute clusters (Niagara, Mist, Trillium,... ) are not and will not be mounted.  You'll need to copy over any files that you need to use on the Teach cluster.&lt;br /&gt;
* There is no $SCRATCH.  You can do all your work on $HOME, which is writable from compute nodes&lt;br /&gt;
* The software stack is the one supplied by the Alliance.  There is no need to load 'CCEnv' to get them.&lt;br /&gt;
* But as before, if you're missing a module, we can still install it for you.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- Do this later&lt;br /&gt;
== Running Jupyter on a Teach Compute Node ==  &lt;br /&gt;
&lt;br /&gt;
1. To be able to run Jupyter on a compute node, you must first (a) install it inside a virtual environment, (b) enable a way for jupyter to seemingly write to a specific directory on $HOME, and (c) create a little helper script called &amp;lt;tt&amp;gt;notebook.sh&amp;lt;/tt&amp;gt; that will be used to start the jupyter server in step 2.  These are the commands that you should use for the installation (which you should do only once, on the Teach login node):&lt;br /&gt;
&lt;br /&gt;
(a) Create virtual env&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
$ module load python/3.9.10&lt;br /&gt;
$ virtualenv --system-site-packages $HOME/.virtualenvs/jupteach&lt;br /&gt;
$ source $HOME/.virtualenvs/jupteach/bin/activate&lt;br /&gt;
$ pip install jupyter jupyterlab&lt;br /&gt;
$ deactivate&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
You can choose another directory than &amp;lt;tt&amp;gt;$HOME/.virtualenvs/jupteach&amp;lt;/tt&amp;gt; for where to create the virtual environment, but you need to be consistent and use the same directory everywhere below.&lt;br /&gt;
&lt;br /&gt;
(b) Make a writable 'runtime' directory for Jupyter. &lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
$ mkdir -p $HOME/.local/share/jupyter/runtime &lt;br /&gt;
$ mv -f $HOME/.local/share/jupyter/runtime $SCRATCH/jupyter_runtime || mkdir $SCRATCH/jupyter_runtime&lt;br /&gt;
$ ln -sT $SCRATCH/jupyter_runtime $HOME/.local/share/jupyter/runtime&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
(c) Create a launch script to use on the compute nodes:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
$ cat &amp;gt; $HOME/.virtualenvs/jupteach/bin/notebook.sh &amp;lt;&amp;lt;EOF&lt;br /&gt;
#!/bin/bash&lt;br /&gt;
source \$HOME/.virtualenvs/jupteach/bin/activate&lt;br /&gt;
export XDG_DATA_HOME=\$SCRATCH/.share&lt;br /&gt;
export XDG_CACHE_HOME=\$SCRATCH/.cache&lt;br /&gt;
export XDG_CONFIG_HOME=\$SCRATCH/.config&lt;br /&gt;
export XDG_RUNTIME_DIR=\$SCRATCH/.runtime&lt;br /&gt;
export JUPYTER_CONFIG_DIR=\$SCRATCH/.config/.jupyter&lt;br /&gt;
jupyter \${1:-notebook} --ip \$(hostname -f) --no-browser --notebook-dir=\$PWD&lt;br /&gt;
EOF&lt;br /&gt;
$ chmod +x  $HOME/.virtualenvs/jupteach/bin/notebook.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2. To run the jupyter server on a compute node, start an interactive session with the &amp;lt;tt&amp;gt;debugjob&amp;lt;/tt&amp;gt; command and then launch the jupyter server:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
$ debugjob -n 16                    # use less if you need less cores.&lt;br /&gt;
$ cd $SCRATCH                       # $HOME is read-only, so move to $SCRATCH&lt;br /&gt;
$ $HOME/.virtualenvs/jupteach/bin/notebook.sh  # add the argument &amp;quot;lab&amp;quot; to start with the jupyter lab&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
Make sure you note down (a) the name of the compute node that you got allocated (they start with &amp;quot;&amp;lt;tt&amp;gt;teach&amp;lt;/tt&amp;gt;&amp;quot; followed by a 2-digit number), and (b) the number following the compute nodes after the colon (usually this is 8888, but it can be another, higher number); this is the PORT  and (c) the last URL that the notebook.sh tells you to use to connect.&lt;br /&gt;
&lt;br /&gt;
4. To connect to this jupyter server running on a teach compute node, which is not accessible from the internet, in a different terminal on you own computer, you must reconnect to the Teach cluster with a port-forwarding&lt;br /&gt;
tunnel to the compute node on which jupyter is running:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
$ ssh -LPORT:teachXX:PORT -o ControlMaster=no USERNAME@teach.scinet.utoronto.ca  -N&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;tt&amp;gt;teachXX&amp;lt;/tt&amp;gt; is to be replaced by the name of the compute node (point (a) above), PORT is to be replaced by the port number that notebook.sh showed, and &amp;lt;tt&amp;gt;USERNAME&amp;lt;/t&amp;gt; should be your teach account username. This command should just &amp;quot;hang&amp;quot; there, it only serves to forward port number PORT (usually 8888) to port PORT (usually 8888) on the compute node.  &lt;br /&gt;
&lt;br /&gt;
Finally, point your browser to the URL that the &amp;lt;tt&amp;gt;notebook.sh&amp;lt;/tt&amp;gt; command printed out (point (b) above), i.e., the one with 127.0.0.1 in it.&lt;br /&gt;
--&amp;gt;&lt;/div&gt;</summary>
		<author><name>Rzon</name></author>
	</entry>
	<entry>
		<id>https://docs.scinet.utoronto.ca/index.php?title=Teach&amp;diff=7514</id>
		<title>Teach</title>
		<link rel="alternate" type="text/html" href="https://docs.scinet.utoronto.ca/index.php?title=Teach&amp;diff=7514"/>
		<updated>2026-02-13T17:22:43Z</updated>

		<summary type="html">&lt;p&gt;Rzon: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Infobox Computer&lt;br /&gt;
|image=[[Image:Ibm_idataplex_dx360_m4.jpg|center|300px|thumb]] &lt;br /&gt;
|name=Teach Cluster &lt;br /&gt;
|installed=(orig Mar 2020), Feb 2025&lt;br /&gt;
|operatingsystem= Linux (Rocky 9.6)&lt;br /&gt;
|loginnode=teach-login01&lt;br /&gt;
|nnodes=64 &lt;br /&gt;
|rampernode=188 GiB / 202 GB &lt;br /&gt;
|corespernode=40 &lt;br /&gt;
|interconnect=Infiniband (EDR)&lt;br /&gt;
|vendorcompilers=gcc,intel&lt;br /&gt;
|queuetype=slurm&lt;br /&gt;
}}&lt;br /&gt;
&lt;br /&gt;
== Teaching Cluster ==&lt;br /&gt;
&lt;br /&gt;
The Teach cluster is a 2560-core cluster provided for teaching purposes.  It has currently no GPU capability.&lt;br /&gt;
It is configured similarly to Trillium, however it uses hardware repurposed from its predecessor, Niagara.  This system should not be used for production work as the queuing policies are designed to provide fast job turnover and limit the amount of resources one person can use at a time.     Questions about its use or problems should be sent to '''support@scinet.utoronto.ca'''.&lt;br /&gt;
&lt;br /&gt;
== Specifications==&lt;br /&gt;
&lt;br /&gt;
This cluster currently consists of 64 repurposed x86_64 nodes each with 40 cores (from two 20 core Intel CascadeLake CPUs) running at 2.5GHz with 188GB of RAM per node. &lt;br /&gt;
The nodes are interconnected with 1:1 non-blocking EDR Infiniband for MPI communications, and disk I/O to a separate view of the VAST file system.  In total, this cluster contains 2560 cores.&lt;br /&gt;
&lt;br /&gt;
== Login/Devel Node ==&lt;br /&gt;
&lt;br /&gt;
Teach runs Rocky Linux 9.  You will need to be somewhat familiar with Linux systems to work on Teach.  If you are not, it will be worth your time to review our [https://education.scinet.utoronto.ca/tag/index.php?tag=SCMP101 Introduction to Linux Shell] class.&lt;br /&gt;
&lt;br /&gt;
As with all SciNet and {{Alliance}} systems, access to Teach is done via [[SSH]] (secure shell) only.  Open a terminal window (e.g. using [https://docs.alliancecan.ca/wiki/Connecting_with_PuTTY PuTTY] or  [https://docs.alliancecan.ca/wiki/Connecting_with_MobaXTerm MobaXTerm] on Windows), and type&lt;br /&gt;
 ssh -Y USERNAME@teach.scinet.utoronto.ca&lt;br /&gt;
This will bring directly to the command line of '''&amp;lt;tt&amp;gt;teach-login01&amp;lt;/tt&amp;gt;''' or '''&amp;lt;tt&amp;gt;teach-login02&amp;lt;/tt&amp;gt;''', which are the gateway/devel nodes for this cluster.  &lt;br /&gt;
On these nodes, you can compile, do short tests, and submit your jobs to the queue.&lt;br /&gt;
&lt;br /&gt;
The first time you login to Teach cluster, please make sure to check if the login node ssh key fingerprint&lt;br /&gt;
matches. [[Teach_fingerprints | See here how]].&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The login nodes are shared between students of a          &lt;br /&gt;
number of different courses. Use this node to develop and compile                  &lt;br /&gt;
code, to run short tests, and to submit computations to the scheduler (see below).&lt;br /&gt;
&lt;br /&gt;
Note that access to the Teach cluster is restricted to temporary accounts that start with the prefix '''lcl_uot''' + the course code + '''s''', and a number.  Passwords for these accounts can be changed on the [https://portal.scinet.utoronto.ca/portaluserlogin SciNet user portal]. On the same site, users can upload a public ssh key if they want to connect using ssh keys.&lt;br /&gt;
&lt;br /&gt;
== Software Modules ==&lt;br /&gt;
&lt;br /&gt;
Other than essentials, all installed software is made available [[Using_modules | using module commands]]. These modules set environment variables (PATH, etc.), allowing multiple, conflicting versions of a given package to be available.  A detailed explanation of the module system can be [[Using_modules | found on the modules page]].  &lt;br /&gt;
&lt;br /&gt;
The Teach cluster makes the same [https://docs.alliancecan.ca/wiki/Available_software modules available] as on the [https://docs.alliancecan.ca General Purpose clusters of the Digital Research Alliance of Canada], with one caveat.  On Teach, by default, only the &amp;quot;gentoo&amp;quot; module is loaded, which provides basic OS-level functionality.  &lt;br /&gt;
&lt;br /&gt;
Common module subcommands are:&lt;br /&gt;
&lt;br /&gt;
* &amp;lt;code&amp;gt;module load &amp;lt;module-name&amp;gt;&amp;lt;/code&amp;gt;: load the default version of a particular software.&lt;br /&gt;
* &amp;lt;code&amp;gt;module load &amp;lt;module-name&amp;gt;/&amp;lt;module-version&amp;gt;&amp;lt;/code&amp;gt;: load a specific version of a particular software.&lt;br /&gt;
* &amp;lt;code&amp;gt;module purge&amp;lt;/code&amp;gt;: unload all currently loaded modules.&lt;br /&gt;
* &amp;lt;code&amp;gt;module spider&amp;lt;/code&amp;gt; (or &amp;lt;code&amp;gt;module spider &amp;lt;module-name&amp;gt;&amp;lt;/code&amp;gt;): list available software packages.&lt;br /&gt;
* &amp;lt;code&amp;gt;module avail&amp;lt;/code&amp;gt;: list loadable software packages.&lt;br /&gt;
* &amp;lt;code&amp;gt;module list&amp;lt;/code&amp;gt;: list loaded modules.&lt;br /&gt;
&lt;br /&gt;
For example, to make the GNU compilers (gcc, g++ and gfortran) available, you should type&lt;br /&gt;
&lt;br /&gt;
 module load gcc&lt;br /&gt;
&lt;br /&gt;
while the Intel compilers (icc, icpc and ifort) can be loaded by&lt;br /&gt;
&lt;br /&gt;
 module load intel&lt;br /&gt;
&lt;br /&gt;
To get the default modules that are loaded on the General Purpose clusters, you can load the &amp;quot;StdEnv&amp;quot; module. &lt;br /&gt;
&lt;br /&gt;
Along with modifying common environment variables, such as the PATH, these modules also create an '''EBROOT&amp;lt;MODULENAME&amp;gt;''' environment variable, which can be used to access commonly needed software directories, such as /include and /lib.&lt;br /&gt;
&lt;br /&gt;
There are handy abbreviations for the module commands. &amp;lt;code&amp;gt;ml&amp;lt;/code&amp;gt; is the same as &amp;lt;code&amp;gt;module list&amp;lt;/code&amp;gt;, and &amp;lt;code&amp;gt;ml &amp;lt;module-name&amp;gt;&amp;lt;/code&amp;gt; is the same as &amp;lt;code&amp;gt;module load &amp;lt;module-name&amp;gt;&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
A list of available software modules can be on [https://docs.alliancecan.ca/wiki/Available_software found on this page].&lt;br /&gt;
&lt;br /&gt;
There are a few addition modules available as well, and more can be made available upon demand of the course instructors.  Currently, the only additional modules are:&lt;br /&gt;
&lt;br /&gt;
 catch2/3.3.1      - A C++ test framework for unit-tests, TDD and BDD using C++14 and later.&lt;br /&gt;
 misopy/0.5.2      - A probabilistic framework to analysize RNA-Seq data.&lt;br /&gt;
 palemoon/33.6.0.1 - A web browser&lt;br /&gt;
&lt;br /&gt;
== Interactive jobs ==&lt;br /&gt;
&lt;br /&gt;
For a interactive sessions on a compute node of the teach cluster that give access to non-shared resources, use the 'debugjob' command. &lt;br /&gt;
 teach01:~$ debugjob -n C&lt;br /&gt;
where C is the number of cores. An interactive session defaults to four hours when using at most one node (C&amp;lt;=40), and becomes 60 minutes when using four nodes (i.e., 120&amp;lt;C&amp;lt;=160), which is the maximum number of nodes allowed for an interactive session by debugjob.&lt;br /&gt;
                                                             &lt;br /&gt;
For a short interactive sessions on a dedicated compute node of the teach cluster, use the 'debugjob' command as follows: &lt;br /&gt;
 teach01:~$ debugjob N&lt;br /&gt;
where N is the number of nodes.  On the Teach cluster, this is equivalent to &amp;lt;tt&amp;gt;debugjob -n 40*N &amp;lt;/tt&amp;gt;. The positive integer number &amp;lt;tt&amp;gt;N&amp;lt;/tt&amp;gt; can at most be 4.&lt;br /&gt;
&lt;br /&gt;
If no arguments are given to &amp;lt;tt&amp;gt;debugjob&amp;lt;/tt&amp;gt;, it allocates a single core on a Teach compute node.&lt;br /&gt;
&lt;br /&gt;
There are limits on the resources you can get with a debugjob, and how long you can get them.  No debugjob can run longer than four hours or use more than 160 cores, and each user can only run one at a time.  For longer computations, jobs must be submitted to the scheduler.&lt;br /&gt;
&lt;br /&gt;
== Submit a Job ==&lt;br /&gt;
&lt;br /&gt;
Teach uses SLURM as its job scheduler.  More-advanced details of how to interact with the scheduler can be found on the [[Slurm | Slurm page]].&lt;br /&gt;
&lt;br /&gt;
You submit jobs from a login node by passing a script to the sbatch command:&lt;br /&gt;
&lt;br /&gt;
 teach-login01:~$ sbatch jobscript.sh&lt;br /&gt;
&lt;br /&gt;
This puts the job in the queue. It will run on the compute nodes in due course.&lt;br /&gt;
&lt;br /&gt;
Note:&lt;br /&gt;
* Make sure to adjust accordingly the flags --ntasks-per-node or --ntasks together with --nodes for the examples found at [[Slurm | Slurm page]]. &lt;br /&gt;
* The current slurm configuration of the teach cluster allocates compute resources by core as opposed to by node. That means your tasks might land on nodes that have other jobs running, i.e. they might share the node. If you want to avoid that, make sure to add the following directive in your submitting script: #SBATCH --exclusive. This forces your job to use the compute nodes exclusively.&lt;br /&gt;
* The maximum wall time is currently set to 4 hours.&lt;br /&gt;
* There are 2 queues available: Compute queue and debug queue. Their usage limits are listed on the table below.&lt;br /&gt;
&lt;br /&gt;
== Limits ==&lt;br /&gt;
There are limits to the size and duration of your jobs, the number of jobs you can run and the number of jobs you can have queued. It also matters in which 'partition' the jobs runs. 'Partitions' are SLURM-speak for use cases. You specify the partition with the -p parameter to sbatch or salloc, but if you do not specify one, your job will run in the compute partition, which is the most common case.&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!Usage&lt;br /&gt;
!Partition&lt;br /&gt;
!Running jobs&lt;br /&gt;
!Submitted jobs (incl. running)&lt;br /&gt;
!Min. size of jobs&lt;br /&gt;
!Max. size of jobs&lt;br /&gt;
!Min. walltime&lt;br /&gt;
!Max. walltime &lt;br /&gt;
|-&lt;br /&gt;
|Interactive testing or troubleshooting || debug || 1 || 1 || 1 core || 4 nodes (160 cores)|| N/A || 4 hours&lt;br /&gt;
|-&lt;br /&gt;
|Compute jobs ||compute || 1 || 12 || 1 core || 4 nodes (160 cores)|| 15 minutes || 4 hours&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
Within these limits, jobs may still have to wait in the queue.  Although there are no allocations on the teach cluster, the waiting time still depends on many factors, such as the number of nodes and the wall time, how many other jobs are waiting in the queue, and whether a job can fill an otherwise unused spot in the schedule.&lt;br /&gt;
&lt;br /&gt;
== Main changes from Teach's predecessor ==&lt;br /&gt;
&lt;br /&gt;
This section is intended for instructors that may have use SciNet's previous Teach cluster.&lt;br /&gt;
&lt;br /&gt;
This Teach cluster is setup differently from its predecessor.  &lt;br /&gt;
&lt;br /&gt;
Although the cluster is once again called ''Teach'' and you connect with to teach.scinet.utoronto.ca as before, the system is setup differently from the previous Teach cluster in the following ways:&lt;br /&gt;
&lt;br /&gt;
* There are now 2 dedicated login nodes, teach-login01 and teach-login02.  &lt;br /&gt;
* The ssh fingerprints for these login nodes can be found on [[Teach_fingerprints]].&lt;br /&gt;
* The compute nodes have 40 cores.  As before, you can request jobs by number of cores.&lt;br /&gt;
* Only temporary lcl_uot.... accounts can log in.&lt;br /&gt;
* Only the home directories of those account are mounted. &lt;br /&gt;
* In particular, the file systems from the other SciNet compute clusters (Niagara, Mist, Trillium,... ) are not and will not be mounted.  You'll need to copy over any files that you need to use on the Teach cluster.&lt;br /&gt;
* There is no $SCRATCH.  You can do all your work on $HOME, which is writable from compute nodes&lt;br /&gt;
* The software stack is the one supplied by the Alliance.  There is no need to load 'CCEnv' to get them.&lt;br /&gt;
* But as before, if you're missing a module, we can still install it for you.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- Do this later&lt;br /&gt;
== Running Jupyter on a Teach Compute Node ==  &lt;br /&gt;
&lt;br /&gt;
1. To be able to run Jupyter on a compute node, you must first (a) install it inside a virtual environment, (b) enable a way for jupyter to seemingly write to a specific directory on $HOME, and (c) create a little helper script called &amp;lt;tt&amp;gt;notebook.sh&amp;lt;/tt&amp;gt; that will be used to start the jupyter server in step 2.  These are the commands that you should use for the installation (which you should do only once, on the Teach login node):&lt;br /&gt;
&lt;br /&gt;
(a) Create virtual env&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
$ module load python/3.9.10&lt;br /&gt;
$ virtualenv --system-site-packages $HOME/.virtualenvs/jupteach&lt;br /&gt;
$ source $HOME/.virtualenvs/jupteach/bin/activate&lt;br /&gt;
$ pip install jupyter jupyterlab&lt;br /&gt;
$ deactivate&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
You can choose another directory than &amp;lt;tt&amp;gt;$HOME/.virtualenvs/jupteach&amp;lt;/tt&amp;gt; for where to create the virtual environment, but you need to be consistent and use the same directory everywhere below.&lt;br /&gt;
&lt;br /&gt;
(b) Make a writable 'runtime' directory for Jupyter. &lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
$ mkdir -p $HOME/.local/share/jupyter/runtime &lt;br /&gt;
$ mv -f $HOME/.local/share/jupyter/runtime $SCRATCH/jupyter_runtime || mkdir $SCRATCH/jupyter_runtime&lt;br /&gt;
$ ln -sT $SCRATCH/jupyter_runtime $HOME/.local/share/jupyter/runtime&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
(c) Create a launch script to use on the compute nodes:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
$ cat &amp;gt; $HOME/.virtualenvs/jupteach/bin/notebook.sh &amp;lt;&amp;lt;EOF&lt;br /&gt;
#!/bin/bash&lt;br /&gt;
source \$HOME/.virtualenvs/jupteach/bin/activate&lt;br /&gt;
export XDG_DATA_HOME=\$SCRATCH/.share&lt;br /&gt;
export XDG_CACHE_HOME=\$SCRATCH/.cache&lt;br /&gt;
export XDG_CONFIG_HOME=\$SCRATCH/.config&lt;br /&gt;
export XDG_RUNTIME_DIR=\$SCRATCH/.runtime&lt;br /&gt;
export JUPYTER_CONFIG_DIR=\$SCRATCH/.config/.jupyter&lt;br /&gt;
jupyter \${1:-notebook} --ip \$(hostname -f) --no-browser --notebook-dir=\$PWD&lt;br /&gt;
EOF&lt;br /&gt;
$ chmod +x  $HOME/.virtualenvs/jupteach/bin/notebook.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2. To run the jupyter server on a compute node, start an interactive session with the &amp;lt;tt&amp;gt;debugjob&amp;lt;/tt&amp;gt; command and then launch the jupyter server:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
$ debugjob -n 16                    # use less if you need less cores.&lt;br /&gt;
$ cd $SCRATCH                       # $HOME is read-only, so move to $SCRATCH&lt;br /&gt;
$ $HOME/.virtualenvs/jupteach/bin/notebook.sh  # add the argument &amp;quot;lab&amp;quot; to start with the jupyter lab&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
Make sure you note down (a) the name of the compute node that you got allocated (they start with &amp;quot;&amp;lt;tt&amp;gt;teach&amp;lt;/tt&amp;gt;&amp;quot; followed by a 2-digit number), and (b) the number following the compute nodes after the colon (usually this is 8888, but it can be another, higher number); this is the PORT  and (c) the last URL that the notebook.sh tells you to use to connect.&lt;br /&gt;
&lt;br /&gt;
4. To connect to this jupyter server running on a teach compute node, which is not accessible from the internet, in a different terminal on you own computer, you must reconnect to the Teach cluster with a port-forwarding&lt;br /&gt;
tunnel to the compute node on which jupyter is running:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
$ ssh -LPORT:teachXX:PORT -o ControlMaster=no USERNAME@teach.scinet.utoronto.ca  -N&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;tt&amp;gt;teachXX&amp;lt;/tt&amp;gt; is to be replaced by the name of the compute node (point (a) above), PORT is to be replaced by the port number that notebook.sh showed, and &amp;lt;tt&amp;gt;USERNAME&amp;lt;/t&amp;gt; should be your teach account username. This command should just &amp;quot;hang&amp;quot; there, it only serves to forward port number PORT (usually 8888) to port PORT (usually 8888) on the compute node.  &lt;br /&gt;
&lt;br /&gt;
Finally, point your browser to the URL that the &amp;lt;tt&amp;gt;notebook.sh&amp;lt;/tt&amp;gt; command printed out (point (b) above), i.e., the one with 127.0.0.1 in it.&lt;br /&gt;
--&amp;gt;&lt;/div&gt;</summary>
		<author><name>Rzon</name></author>
	</entry>
	<entry>
		<id>https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7499</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7499"/>
		<updated>2026-01-29T15:47:33Z</updated>

		<summary type="html">&lt;p&gt;Rzon: /* System Status */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;__NOTOC__&lt;br /&gt;
{| style=&amp;quot;border-spacing:10px; width: 95%&amp;quot;&lt;br /&gt;
| style=&amp;quot;padding:1em; padding-top:.1em; border:2px solid #0645ad; background-color:#f6f6f6; border-radius:7px&amp;quot;|&lt;br /&gt;
&lt;br /&gt;
==System Status==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- Use &amp;quot;Up&amp;quot;, &amp;quot;Partial&amp;quot; or &amp;quot;Down&amp;quot;; these are templates. --&amp;gt;&lt;br /&gt;
{|style=&amp;quot;width:100%&amp;quot; &lt;br /&gt;
|{{Down3 | Trillium|https://docs.alliancecan.ca/wiki/Trillium_Quickstart}}&lt;br /&gt;
|{{Down3 | OnDemand|https://docs.alliancecan.ca/wiki/Trillium_Open_OnDemand_Quickstart}}&lt;br /&gt;
|{{Up | Globus |Globus}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Up | HPSS|HPSS}}&lt;br /&gt;
|{{Up | Balam|Balam}}&lt;br /&gt;
|{{Up | S4H | S4H}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Up | Teach|Teach}}&lt;br /&gt;
|{{Up3 | File system|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Storage}}&lt;br /&gt;
|{{Up3 | External Network|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Logging_in}} &lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 29, 2026, 10:00 am:''' There was a power glitch at the data centre overnight. The login nodes are accessible but the compute nodes are down.  &lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 16, 2026, 11:00 pm:''' HPSS is back online, and accessible via alliancecan#hpss Globus endpoint. &lt;br /&gt;
&lt;br /&gt;
'''Thu Jan 15, 2026, 10:00 pm:''' HPSS will undergo maintenance on Friday morning, Jan/16/2025, , including alliancecan#hpss Globus endpoint &lt;br /&gt;
&lt;br /&gt;
'''Tue Jan 6, 2026, 10:15 am:''' OnDemand has been fixed and is working again.&lt;br /&gt;
&lt;br /&gt;
'''Mon Jan 5, 2026, 9:00 pm:''' The authentication mechanism of OnDemand is not working.&lt;br /&gt;
&lt;br /&gt;
'''Wed Dec 31, 2025, 12:40 pm:''' We believe the problem has now been resolved.  Please let us know if you still experience login problems or aborted jobs.&lt;br /&gt;
&lt;br /&gt;
'''Tue Dec 30, 2025, 2:10 pm:''' We are experiencing problems with authentication, resulting in failed logins, OOD errors, and aborted jobs (with &amp;quot;prolog error&amp;quot;).  Please bear with us, as we are very short-staffed during the holiday break.  We will post updates here.&lt;br /&gt;
&lt;br /&gt;
'''Tue Dec 3, 2025, 11:30 am:''' Open OnDemand is fully operational again.&lt;br /&gt;
&lt;br /&gt;
'''Sat Nov 29, 2025, 00:40 am:''' There has been a problem with the water chiller. Some systems are offline.&lt;br /&gt;
&lt;br /&gt;
'''Wed Nov 5, 2025, 12:55 pm:''' Balam is back online.&lt;br /&gt;
&lt;br /&gt;
'''Wed Nov 5, 2025, 10:00 am:''' Open OnDemand is back online.&lt;br /&gt;
&lt;br /&gt;
'''Tue Nov 4, 2025, 11:00 pm:''' Most of the work is done, data movers, Globus, and HPSS are back online. Remaining services will be worked on tomorrow.&lt;br /&gt;
&lt;br /&gt;
'''Tue Nov 4, 2025, 8:30 am:''' Scheduled network maintenance. Trillium cluster is *not* affected.&lt;br /&gt;
&lt;br /&gt;
'''Tue Oct 21, 2025, 17:30 am:''' Balam maintenance finished.&lt;br /&gt;
&lt;br /&gt;
'''Tue Oct 21, 2025, 7:00 am:''' Balam maintenance day.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 15, 2025, 3:55 pm:''' Trillium inbound connections through trillium.alliancecan.ca or trillium.scinet.utoronto.ca are working again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 15, 2025, 3:05 pm:''' Trillium is experiencing external network issues for both incoming traffic. Please try: ssh USERNAME@tri-login01.scinet.utoronto.ca in the meantime.&lt;br /&gt;
 &lt;br /&gt;
'''Thu Oct 06, 2025, 8:00 pm:''' HPSS is fully functional. You may submit archive jobs from trillium login nodes, datamovers and robots.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 03, 2025, 6:30 pm:''' HPSS is back online, and already accessible via alliancecan#hpss Globus endpoint. Directory tree now follows the other Alliance clusters. We're still working on job submission via Slurm&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 01, 2025, 0:00 am:''' Niagara compute nodes are now unavailable for regular users. The login nodes will remain available for a while to allow a few last data transfers, although transfers from the Niagara file systems to Trillium are best done on nia-dm1.scinet.utoronto.ca.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 01, 2025, 9:30 am:''' HPSS is down for scheduled maintenance, including alliancecan#hpss Globus endpoint&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--  When removing system status entries, please archive them to: --&amp;gt;&lt;br /&gt;
[[Previous messages]]&lt;br /&gt;
&lt;br /&gt;
{|style=&amp;quot;border-spacing: 10px;width: 100%&amp;quot;&lt;br /&gt;
|valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== QuickStart Guides ==&lt;br /&gt;
* [https://docs.alliancecan.ca/wiki/Trillium_Quickstart Trillium Quickstart]&lt;br /&gt;
* [[Niagara Quickstart]]&lt;br /&gt;
* [[HPSS | HPSS archival storage]]&lt;br /&gt;
* [[Teach|Teach cluster]]&lt;br /&gt;
* [[FAQ | FAQ (frequently asked questions)]]&lt;br /&gt;
* [[Acknowledging SciNet]]&lt;br /&gt;
| valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== Tutorials, Manuals, etc. ==&lt;br /&gt;
* [https://education.scinet.utoronto.ca SciNet education material]&lt;br /&gt;
* [https://www.youtube.com/c/SciNetHPCattheUniversityofToronto SciNet's YouTube channel]&lt;br /&gt;
* [[Modules specific to Niagara|Software Modules specific to Niagara]] &lt;br /&gt;
* [[Modules for Mist]] &lt;br /&gt;
* [[Commercial software]]&lt;br /&gt;
* [[Burst Buffer]]&lt;br /&gt;
* [[SSH#SSH Keys|SSH keys]]&lt;br /&gt;
* [[SSH Tunneling]]&lt;br /&gt;
* [[Visualization]]&lt;br /&gt;
* [[Running Serial Jobs on Niagara]]&lt;br /&gt;
* [[Jupyter Hub]]&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Rzon</name></author>
	</entry>
	<entry>
		<id>https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7478</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7478"/>
		<updated>2026-01-07T19:49:53Z</updated>

		<summary type="html">&lt;p&gt;Rzon: /* System Status */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;__NOTOC__&lt;br /&gt;
{| style=&amp;quot;border-spacing:10px; width: 95%&amp;quot;&lt;br /&gt;
| style=&amp;quot;padding:1em; padding-top:.1em; border:2px solid #0645ad; background-color:#f6f6f6; border-radius:7px&amp;quot;|&lt;br /&gt;
&lt;br /&gt;
==System Status==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- Use &amp;quot;Up&amp;quot;, &amp;quot;Partial&amp;quot; or &amp;quot;Down&amp;quot;; these are templates. --&amp;gt;&lt;br /&gt;
{|style=&amp;quot;width:100%&amp;quot; &lt;br /&gt;
|{{Up3 | Trillium|https://docs.alliancecan.ca/wiki/Trillium_Quickstart}}&lt;br /&gt;
|{{Up3 | OnDemand|https://docs.alliancecan.ca/wiki/Trillium_Open_OnDemand_Quickstart}}&lt;br /&gt;
|{{Up | Globus |Globus}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Up | HPSS|HPSS}}&lt;br /&gt;
|{{Up | Balam|Balam}}&lt;br /&gt;
|{{Up | S4H | S4H}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Up | Teach|Teach}}&lt;br /&gt;
|{{Up3 | File system|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Storage}}&lt;br /&gt;
|{{Up3 | External Network|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Logging_in}} &lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Tue Jan 6, 2026, 10:15 am:''' OnDemand has been fixed and is working again.&lt;br /&gt;
&lt;br /&gt;
'''Mon Jan 5, 2026, 9:00 pm:''' The authentication mechanism of OnDemand is not working.&lt;br /&gt;
&lt;br /&gt;
'''Wed Dec 31, 2025, 12:40 pm:''' We believe the problem has now been resolved.  Please let us know if you still experience login problems or aborted jobs.&lt;br /&gt;
&lt;br /&gt;
'''Tue Dec 30, 2025, 2:10 pm:''' We are experiencing problems with authentication, resulting in failed logins, OOD errors, and aborted jobs (with &amp;quot;prolog error&amp;quot;).  Please bear with us, as we are very short-staffed during the holiday break.  We will post updates here.&lt;br /&gt;
&lt;br /&gt;
'''Tue Dec 3, 2025, 11:30 am:''' Open OnDemand is fully operational again.&lt;br /&gt;
&lt;br /&gt;
'''Sat Nov 29, 2025, 00:40 am:''' There has been a problem with the water chiller. Some systems are offline.&lt;br /&gt;
&lt;br /&gt;
'''Wed Nov 5, 2025, 12:55 pm:''' Balam is back online.&lt;br /&gt;
&lt;br /&gt;
'''Wed Nov 5, 2025, 10:00 am:''' Open OnDemand is back online.&lt;br /&gt;
&lt;br /&gt;
'''Tue Nov 4, 2025, 11:00 pm:''' Most of the work is done, data movers, Globus, and HPSS are back online. Remaining services will be worked on tomorrow.&lt;br /&gt;
&lt;br /&gt;
'''Tue Nov 4, 2025, 8:30 am:''' Scheduled network maintenance. Trillium cluster is *not* affected.&lt;br /&gt;
&lt;br /&gt;
'''Tue Oct 21, 2025, 17:30 am:''' Balam maintenance finished.&lt;br /&gt;
&lt;br /&gt;
'''Tue Oct 21, 2025, 7:00 am:''' Balam maintenance day.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 15, 2025, 3:55 pm:''' Trillium inbound connections through trillium.alliancecan.ca or trillium.scinet.utoronto.ca are working again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 15, 2025, 3:05 pm:''' Trillium is experiencing external network issues for both incoming traffic. Please try: ssh USERNAME@tri-login01.scinet.utoronto.ca in the meantime.&lt;br /&gt;
 &lt;br /&gt;
'''Thu Oct 06, 2025, 8:00 pm:''' HPSS is fully functional. You may submit archive jobs from trillium login nodes, datamovers and robots.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 03, 2025, 6:30 pm:''' HPSS is back online, and already accessible via alliancecan#hpss Globus endpoint. Directory tree now follows the other Alliance clusters. We're still working on job submission via Slurm&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 01, 2025, 0:00 am:''' Niagara compute nodes are now unavailable for regular users. The login nodes will remain available for a while to allow a few last data transfers, although transfers from the Niagara file systems to Trillium are best done on nia-dm1.scinet.utoronto.ca.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 01, 2025, 9:30 am:''' HPSS is down for scheduled maintenance, including alliancecan#hpss Globus endpoint&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--  When removing system status entries, please archive them to: --&amp;gt;&lt;br /&gt;
[[Previous messages]]&lt;br /&gt;
&lt;br /&gt;
{|style=&amp;quot;border-spacing: 10px;width: 100%&amp;quot;&lt;br /&gt;
|valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== QuickStart Guides ==&lt;br /&gt;
* [https://docs.alliancecan.ca/wiki/Trillium_Quickstart Trillium Quickstart]&lt;br /&gt;
* [[Niagara Quickstart]]&lt;br /&gt;
* [[HPSS | HPSS archival storage]]&lt;br /&gt;
* [[Teach|Teach cluster]]&lt;br /&gt;
* [[FAQ | FAQ (frequently asked questions)]]&lt;br /&gt;
* [[Acknowledging SciNet]]&lt;br /&gt;
| valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== Tutorials, Manuals, etc. ==&lt;br /&gt;
* [https://education.scinet.utoronto.ca SciNet education material]&lt;br /&gt;
* [https://www.youtube.com/c/SciNetHPCattheUniversityofToronto SciNet's YouTube channel]&lt;br /&gt;
* [[Modules specific to Niagara|Software Modules specific to Niagara]] &lt;br /&gt;
* [[Modules for Mist]] &lt;br /&gt;
* [[Commercial software]]&lt;br /&gt;
* [[Burst Buffer]]&lt;br /&gt;
* [[SSH#SSH Keys|SSH keys]]&lt;br /&gt;
* [[SSH Tunneling]]&lt;br /&gt;
* [[Visualization]]&lt;br /&gt;
* [[Running Serial Jobs on Niagara]]&lt;br /&gt;
* [[Jupyter Hub]]&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Rzon</name></author>
	</entry>
	<entry>
		<id>https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7469</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://docs.scinet.utoronto.ca/index.php?title=Main_Page&amp;diff=7469"/>
		<updated>2026-01-06T02:08:16Z</updated>

		<summary type="html">&lt;p&gt;Rzon: /* System Status */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;__NOTOC__&lt;br /&gt;
{| style=&amp;quot;border-spacing:10px; width: 95%&amp;quot;&lt;br /&gt;
| style=&amp;quot;padding:1em; padding-top:.1em; border:2px solid #0645ad; background-color:#f6f6f6; border-radius:7px&amp;quot;|&lt;br /&gt;
&lt;br /&gt;
==System Status==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- Use &amp;quot;Up&amp;quot;, &amp;quot;Partial&amp;quot; or &amp;quot;Down&amp;quot;; these are templates. --&amp;gt;&lt;br /&gt;
{|style=&amp;quot;width:100%&amp;quot; &lt;br /&gt;
|{{Up3 | Trillium|https://docs.alliancecan.ca/wiki/Trillium_Quickstart}}&lt;br /&gt;
|{{Down | OnDemand|Open_OnDemand_Quickstart}}&lt;br /&gt;
|{{Up | Globus |Globus}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Up | HPSS|HPSS}}&lt;br /&gt;
|{{Up | Balam|Balam}}&lt;br /&gt;
|{{Up | S4H | S4H}}&lt;br /&gt;
|-&lt;br /&gt;
|{{Up | Teach|Teach}}&lt;br /&gt;
|{{Up3 | File system|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Storage}}&lt;br /&gt;
|{{Up3 | External Network|https://docs.alliancecan.ca/wiki/Trillium_Quickstart#Logging_in}} &lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Mon Jan 5, 2026, 9:00 pm:''' The authentication mechanism of OnDemand is not working.&lt;br /&gt;
&lt;br /&gt;
'''Wed Dec 31, 2025, 12:40 pm:''' We believe the problem has now been resolved.  Please let us know if you still experience login problems or aborted jobs.&lt;br /&gt;
&lt;br /&gt;
'''Tue Dec 30, 2025, 2:10 pm:''' We are experiencing problems with authentication, resulting in failed logins, OOD errors, and aborted jobs (with &amp;quot;prolog error&amp;quot;).  Please bear with us, as we are very short-staffed during the holiday break.  We will post updates here.&lt;br /&gt;
&lt;br /&gt;
'''Tue Dec 3, 2025, 11:30 am:''' Open OnDemand is fully operational again.&lt;br /&gt;
&lt;br /&gt;
'''Sat Nov 29, 2025, 00:40 am:''' There has been a problem with the water chiller. Some systems are offline.&lt;br /&gt;
&lt;br /&gt;
'''Wed Nov 5, 2025, 12:55 pm:''' Balam is back online.&lt;br /&gt;
&lt;br /&gt;
'''Wed Nov 5, 2025, 10:00 am:''' Open OnDemand is back online.&lt;br /&gt;
&lt;br /&gt;
'''Tue Nov 4, 2025, 11:00 pm:''' Most of the work is done, data movers, Globus, and HPSS are back online. Remaining services will be worked on tomorrow.&lt;br /&gt;
&lt;br /&gt;
'''Tue Nov 4, 2025, 8:30 am:''' Scheduled network maintenance. Trillium cluster is *not* affected.&lt;br /&gt;
&lt;br /&gt;
'''Tue Oct 21, 2025, 17:30 am:''' Balam maintenance finished.&lt;br /&gt;
&lt;br /&gt;
'''Tue Oct 21, 2025, 7:00 am:''' Balam maintenance day.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 15, 2025, 3:55 pm:''' Trillium inbound connections through trillium.alliancecan.ca or trillium.scinet.utoronto.ca are working again.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 15, 2025, 3:05 pm:''' Trillium is experiencing external network issues for both incoming traffic. Please try: ssh USERNAME@tri-login01.scinet.utoronto.ca in the meantime.&lt;br /&gt;
 &lt;br /&gt;
'''Thu Oct 06, 2025, 8:00 pm:''' HPSS is fully functional. You may submit archive jobs from trillium login nodes, datamovers and robots.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 03, 2025, 6:30 pm:''' HPSS is back online, and already accessible via alliancecan#hpss Globus endpoint. Directory tree now follows the other Alliance clusters. We're still working on job submission via Slurm&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 01, 2025, 0:00 am:''' Niagara compute nodes are now unavailable for regular users. The login nodes will remain available for a while to allow a few last data transfers, although transfers from the Niagara file systems to Trillium are best done on nia-dm1.scinet.utoronto.ca.&lt;br /&gt;
&lt;br /&gt;
'''Thu Oct 01, 2025, 9:30 am:''' HPSS is down for scheduled maintenance, including alliancecan#hpss Globus endpoint&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--  When removing system status entries, please archive them to: --&amp;gt;&lt;br /&gt;
[[Previous messages]]&lt;br /&gt;
&lt;br /&gt;
{|style=&amp;quot;border-spacing: 10px;width: 100%&amp;quot;&lt;br /&gt;
|valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== QuickStart Guides ==&lt;br /&gt;
* [https://docs.alliancecan.ca/wiki/Trillium_Quickstart Trillium Quickstart]&lt;br /&gt;
* [[Niagara Quickstart]]&lt;br /&gt;
* [[HPSS | HPSS archival storage]]&lt;br /&gt;
* [[Teach|Teach cluster]]&lt;br /&gt;
* [[FAQ | FAQ (frequently asked questions)]]&lt;br /&gt;
* [[Acknowledging SciNet]]&lt;br /&gt;
| valign=&amp;quot;top&amp;quot; style=&amp;quot;margin: 1em; padding:1em; padding-top:.1em; border:2px solid #000; background-color:#fff; border-radius:7px; width: 49.5%&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
== Tutorials, Manuals, etc. ==&lt;br /&gt;
* [https://education.scinet.utoronto.ca SciNet education material]&lt;br /&gt;
* [https://www.youtube.com/c/SciNetHPCattheUniversityofToronto SciNet's YouTube channel]&lt;br /&gt;
* [[Modules specific to Niagara|Software Modules specific to Niagara]] &lt;br /&gt;
* [[Modules for Mist]] &lt;br /&gt;
* [[Commercial software]]&lt;br /&gt;
* [[Burst Buffer]]&lt;br /&gt;
* [[SSH#SSH Keys|SSH keys]]&lt;br /&gt;
* [[SSH Tunneling]]&lt;br /&gt;
* [[Visualization]]&lt;br /&gt;
* [[Running Serial Jobs on Niagara]]&lt;br /&gt;
* [[Jupyter Hub]]&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Rzon</name></author>
	</entry>
	<entry>
		<id>https://docs.scinet.utoronto.ca/index.php?title=Teach&amp;diff=7466</id>
		<title>Teach</title>
		<link rel="alternate" type="text/html" href="https://docs.scinet.utoronto.ca/index.php?title=Teach&amp;diff=7466"/>
		<updated>2026-01-05T14:31:48Z</updated>

		<summary type="html">&lt;p&gt;Rzon: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Infobox Computer&lt;br /&gt;
|image=[[Image:Ibm_idataplex_dx360_m4.jpg|center|300px|thumb]] &lt;br /&gt;
|name=Teach Cluster &lt;br /&gt;
|installed=(orig Mar 2020), Feb 2025&lt;br /&gt;
|operatingsystem= Linux (Rocky 9.6)&lt;br /&gt;
|loginnode=teach-login01&lt;br /&gt;
|nnodes=64 &lt;br /&gt;
|rampernode=188 GiB / 202 GB &lt;br /&gt;
|corespernode=40 &lt;br /&gt;
|interconnect=Infiniband (EDR)&lt;br /&gt;
|vendorcompilers=gcc,intel&lt;br /&gt;
|queuetype=slurm&lt;br /&gt;
}}&lt;br /&gt;
&lt;br /&gt;
== Teaching Cluster ==&lt;br /&gt;
&lt;br /&gt;
The Teach cluster is a 2560-core cluster provided for teaching purposes.  It has currently no GPU capability.&lt;br /&gt;
It is configured similarly to the coming SciNet production systems Trillium, however it uses hardware repurposed from its predecessor, [[Niagara_Quickstart|Niagara]].  This system should not be used for production work as the queuing policies are designed to provide fast job turnover and limit the amount of resources one person can use at a time.     Questions about its use or problems should be sent to '''support@scinet.utoronto.ca'''.&lt;br /&gt;
&lt;br /&gt;
== Specifications==&lt;br /&gt;
&lt;br /&gt;
This cluster currently consists of 64 repurposed x86_64 nodes each with 40 cores (from two 20 core Intel CascadeLake CPUs) running at 2.5GHz with 188GB of RAM per node. &lt;br /&gt;
The nodes are interconnected with 1:1 non-blocking EDR Infiniband for MPI communications, and disk I/O to a separate view of the VAST file system.  In total, this cluster contains 320 cores, but there are plans to expand it if demand warrants it.&lt;br /&gt;
&lt;br /&gt;
== Login/Devel Node ==&lt;br /&gt;
&lt;br /&gt;
Teach runs Rocky Linux 9.  You will need to be somewhat familiar with Linux systems to work on Teach.  If you are not, it will be worth your time to review our [https://education.scinet.utoronto.ca/tag/index.php?tag=SCMP101 Introduction to Linux Shell] class.&lt;br /&gt;
&lt;br /&gt;
As with all SciNet and {{Alliance}} systems, access to Teach is done via [[SSH]] (secure shell) only.  Open a terminal window (e.g. using [https://docs.alliancecan.ca/wiki/Connecting_with_PuTTY PuTTY] or  [https://docs.alliancecan.ca/wiki/Connecting_with_MobaXTerm MobaXTerm] on Windows), and type&lt;br /&gt;
 ssh -Y USERNAME@teach.scinet.utoronto.ca&lt;br /&gt;
This will bring directly to the command line of '''&amp;lt;tt&amp;gt;teach-login01&amp;lt;/tt&amp;gt;''' or '''&amp;lt;tt&amp;gt;teach-login02&amp;lt;/tt&amp;gt;''', which are the gateway/devel nodes for this cluster.  &lt;br /&gt;
On these nodes, you can compile, do short tests, and submit your jobs to the queue.&lt;br /&gt;
&lt;br /&gt;
The first time you login to Teach cluster, please make sure to check if the login node ssh key fingerprint&lt;br /&gt;
matches. [[Teach_fingerprints | See here how]].&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The login nodes are shared between students of a          &lt;br /&gt;
number of different courses. Use this node to develop and compile                  &lt;br /&gt;
code, to run short tests, and to submit computations to the scheduler (see below).&lt;br /&gt;
&lt;br /&gt;
Note that access to the teach cluster is restricted to temporary accounts that start with the prefix '''lcl_uot''' + the course code + '''s''', and a number.  Passwords for these accounts can be changed on the [https://portal.scinet.utoronto.ca/portaluserlogin SciNet user portal]. On the same site, users can upload a public ssh key if they want to connect using ssh keys.&lt;br /&gt;
&lt;br /&gt;
== Software Modules ==&lt;br /&gt;
&lt;br /&gt;
Other than essentials, all installed software is made available [[Using_modules | using module commands]]. These modules set environment variables (PATH, etc.), allowing multiple, conflicting versions of a given package to be available.  A detailed explanation of the module system can be [[Using_modules | found on the modules page]].  &lt;br /&gt;
&lt;br /&gt;
The Teach cluster makes the same [https://docs.alliancecan.ca/wiki/Available_software modules available] as on the [https://docs.alliancecan.ca General Purpose clusters of the Digital Research Alliance of Canada], with one caveat.  On Teach, by default, only the &amp;quot;gentoo&amp;quot; module is loaded, which provides basic OS-level functionality.  &lt;br /&gt;
&lt;br /&gt;
Common module subcommands are:&lt;br /&gt;
&lt;br /&gt;
* &amp;lt;code&amp;gt;module load &amp;lt;module-name&amp;gt;&amp;lt;/code&amp;gt;: load the default version of a particular software.&lt;br /&gt;
* &amp;lt;code&amp;gt;module load &amp;lt;module-name&amp;gt;/&amp;lt;module-version&amp;gt;&amp;lt;/code&amp;gt;: load a specific version of a particular software.&lt;br /&gt;
* &amp;lt;code&amp;gt;module purge&amp;lt;/code&amp;gt;: unload all currently loaded modules.&lt;br /&gt;
* &amp;lt;code&amp;gt;module spider&amp;lt;/code&amp;gt; (or &amp;lt;code&amp;gt;module spider &amp;lt;module-name&amp;gt;&amp;lt;/code&amp;gt;): list available software packages.&lt;br /&gt;
* &amp;lt;code&amp;gt;module avail&amp;lt;/code&amp;gt;: list loadable software packages.&lt;br /&gt;
* &amp;lt;code&amp;gt;module list&amp;lt;/code&amp;gt;: list loaded modules.&lt;br /&gt;
&lt;br /&gt;
For example, to make the GNU compilers (gcc, g++ and gfortran) available, you should type&lt;br /&gt;
&lt;br /&gt;
 module load gcc&lt;br /&gt;
&lt;br /&gt;
while the Intel compilers (icc, icpc and ifort) can be loaded by&lt;br /&gt;
&lt;br /&gt;
 module load intel&lt;br /&gt;
&lt;br /&gt;
To get the default modules that are loaded on the General Purpose clusters, you can load the &amp;quot;StdEnv&amp;quot; module. &lt;br /&gt;
&lt;br /&gt;
Along with modifying common environment variables, such as the PATH, these modules also create an '''EBROOT&amp;lt;MODULENAME&amp;gt;''' environment variable, which can be used to access commonly needed software directories, such as /include and /lib.&lt;br /&gt;
&lt;br /&gt;
There are handy abbreviations for the module commands. &amp;lt;code&amp;gt;ml&amp;lt;/code&amp;gt; is the same as &amp;lt;code&amp;gt;module list&amp;lt;/code&amp;gt;, and &amp;lt;code&amp;gt;ml &amp;lt;module-name&amp;gt;&amp;lt;/code&amp;gt; is the same as &amp;lt;code&amp;gt;module load &amp;lt;module-name&amp;gt;&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
A list of available software modules can be on [https://docs.alliancecan.ca/wiki/Available_software found on this page].&lt;br /&gt;
&lt;br /&gt;
There are a few addition modules available as well, and more can be made available upon demand of the course instructors.  Currently, the only additional modules are:&lt;br /&gt;
&lt;br /&gt;
 catch2/3.3.1      - A C++ test framework for unit-tests, TDD and BDD using C++14 and later.&lt;br /&gt;
 misopy/0.5.2      - A probabilistic framework to analysize RNA-Seq data.&lt;br /&gt;
 palemoon/33.6.0.1 - A web browser&lt;br /&gt;
&lt;br /&gt;
== Interactive jobs ==&lt;br /&gt;
&lt;br /&gt;
For a interactive sessions on a compute node of the teach cluster that give access to non-shared resources, use the 'debugjob' command. &lt;br /&gt;
 teach01:~$ debugjob -n C&lt;br /&gt;
where C is the number of cores. An interactive session defaults to four hours when using at most one node (C&amp;lt;=40), and becomes 60 minutes when using four nodes (i.e., 120&amp;lt;C&amp;lt;=160), which is the maximum number of nodes allowed for an interactive session by debugjob.&lt;br /&gt;
                                                             &lt;br /&gt;
For a short interactive sessions on a dedicated compute node of the teach cluster, use the 'debugjob' command as follows: &lt;br /&gt;
 teach01:~$ debugjob N&lt;br /&gt;
where N is the number of nodes.  On the Teach cluster, this is equivalent to &amp;lt;tt&amp;gt;debugjob -n 40*N &amp;lt;/tt&amp;gt;. The positive integer number &amp;lt;tt&amp;gt;N&amp;lt;/tt&amp;gt; can at most be 4.&lt;br /&gt;
&lt;br /&gt;
If no arguments are given to &amp;lt;tt&amp;gt;debugjob&amp;lt;/tt&amp;gt;, it allocates a single core on a Teach compute node.&lt;br /&gt;
&lt;br /&gt;
There are limits on the resources you can get with a debugjob, and how long you can get them.  No debugjob can run longer than four hours or use more than 160 cores, and each user can only run one at a time.  For longer computations, jobs must be submitted to the scheduler.&lt;br /&gt;
&lt;br /&gt;
== Submit a Job ==&lt;br /&gt;
&lt;br /&gt;
Teach uses SLURM as its job scheduler.  More-advanced details of how to interact with the scheduler can be found on the [[Slurm | Slurm page]].&lt;br /&gt;
&lt;br /&gt;
You submit jobs from a login node by passing a script to the sbatch command:&lt;br /&gt;
&lt;br /&gt;
 teach-login01:~$ sbatch jobscript.sh&lt;br /&gt;
&lt;br /&gt;
This puts the job in the queue. It will run on the compute nodes in due course.&lt;br /&gt;
&lt;br /&gt;
Note:&lt;br /&gt;
* Make sure to adjust accordingly the flags --ntasks-per-node or --ntasks together with --nodes for the examples found at [[Slurm | Slurm page]]. &lt;br /&gt;
* The current slurm configuration of the teach cluster allocates compute resources by core as opposed to by node. That means your tasks might land on nodes that have other jobs running, i.e. they might share the node. If you want to avoid that, make sure to add the following directive in your submitting script: #SBATCH --exclusive. This forces your job to use the compute nodes exclusively.&lt;br /&gt;
* The maximum wall time is currently set to 4 hours.&lt;br /&gt;
* There are 2 queues available: Compute queue and debug queue. Their usage limits are listed on the table below.&lt;br /&gt;
&lt;br /&gt;
== Limits ==&lt;br /&gt;
There are limits to the size and duration of your jobs, the number of jobs you can run and the number of jobs you can have queued. It also matters in which 'partition' the jobs runs. 'Partitions' are SLURM-speak for use cases. You specify the partition with the -p parameter to sbatch or salloc, but if you do not specify one, your job will run in the compute partition, which is the most common case.&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!Usage&lt;br /&gt;
!Partition&lt;br /&gt;
!Running jobs&lt;br /&gt;
!Submitted jobs (incl. running)&lt;br /&gt;
!Min. size of jobs&lt;br /&gt;
!Max. size of jobs&lt;br /&gt;
!Min. walltime&lt;br /&gt;
!Max. walltime &lt;br /&gt;
|-&lt;br /&gt;
|Interactive testing or troubleshooting || debug || 1 || 1 || 1 core || 4 nodes (160 cores)|| N/A || 4 hours&lt;br /&gt;
|-&lt;br /&gt;
|Compute jobs ||compute || 1 || 12 || 1 core || 4 nodes (160 cores)|| 15 minutes || 4 hours&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
Within these limits, jobs may still have to wait in the queue.  Although there are no allocations on the teach cluster, the waiting time still depends on many factors, such as the number of nodes and the wall time, how many other jobs are waiting in the queue, and whether a job can fill an otherwise unused spot in the schedule.&lt;br /&gt;
&lt;br /&gt;
== Main changes from Teach's predecessor ==&lt;br /&gt;
&lt;br /&gt;
This section is intended for instructors that may have use SciNet's previous Teach cluster.&lt;br /&gt;
&lt;br /&gt;
This Teach cluster is setup differently from its predecessor.  &lt;br /&gt;
&lt;br /&gt;
Although the cluster is once again called ''Teach'' and you connect with to teach.scinet.utoronto.ca as before, the system is setup differently from the previous Teach cluster in the following ways:&lt;br /&gt;
&lt;br /&gt;
* There are now 2 dedicated login nodes, teach-login01 and teach-login02.  &lt;br /&gt;
* The ssh fingerprints for these login nodes can be found on [[Teach_fingerprints]].&lt;br /&gt;
* The compute nodes have 40 cores.  As before, you can request jobs by number of cores.&lt;br /&gt;
* Only temporary lcl_uot.... accounts can log in.&lt;br /&gt;
* Only the home directories of those account are mounted. &lt;br /&gt;
* In particular, the file systems from the other SciNet compute clusters (Niagara, Mist, Trillium,... ) are not and will not be mounted.  You'll need to copy over any files that you need to use on the Teach cluster.&lt;br /&gt;
* There is no $SCRATCH.  You can do all your work on $HOME, which is writable from compute nodes&lt;br /&gt;
* The software stack is the one supplied by the Alliance.  There is no need to load 'CCEnv' to get them.&lt;br /&gt;
* But as before, if you're missing a module, we can still install it for you.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- Do this later&lt;br /&gt;
== Running Jupyter on a Teach Compute Node ==  &lt;br /&gt;
&lt;br /&gt;
1. To be able to run Jupyter on a compute node, you must first (a) install it inside a virtual environment, (b) enable a way for jupyter to seemingly write to a specific directory on $HOME, and (c) create a little helper script called &amp;lt;tt&amp;gt;notebook.sh&amp;lt;/tt&amp;gt; that will be used to start the jupyter server in step 2.  These are the commands that you should use for the installation (which you should do only once, on the Teach login node):&lt;br /&gt;
&lt;br /&gt;
(a) Create virtual env&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
$ module load python/3.9.10&lt;br /&gt;
$ virtualenv --system-site-packages $HOME/.virtualenvs/jupteach&lt;br /&gt;
$ source $HOME/.virtualenvs/jupteach/bin/activate&lt;br /&gt;
$ pip install jupyter jupyterlab&lt;br /&gt;
$ deactivate&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
You can choose another directory than &amp;lt;tt&amp;gt;$HOME/.virtualenvs/jupteach&amp;lt;/tt&amp;gt; for where to create the virtual environment, but you need to be consistent and use the same directory everywhere below.&lt;br /&gt;
&lt;br /&gt;
(b) Make a writable 'runtime' directory for Jupyter. &lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
$ mkdir -p $HOME/.local/share/jupyter/runtime &lt;br /&gt;
$ mv -f $HOME/.local/share/jupyter/runtime $SCRATCH/jupyter_runtime || mkdir $SCRATCH/jupyter_runtime&lt;br /&gt;
$ ln -sT $SCRATCH/jupyter_runtime $HOME/.local/share/jupyter/runtime&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
(c) Create a launch script to use on the compute nodes:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
$ cat &amp;gt; $HOME/.virtualenvs/jupteach/bin/notebook.sh &amp;lt;&amp;lt;EOF&lt;br /&gt;
#!/bin/bash&lt;br /&gt;
source \$HOME/.virtualenvs/jupteach/bin/activate&lt;br /&gt;
export XDG_DATA_HOME=\$SCRATCH/.share&lt;br /&gt;
export XDG_CACHE_HOME=\$SCRATCH/.cache&lt;br /&gt;
export XDG_CONFIG_HOME=\$SCRATCH/.config&lt;br /&gt;
export XDG_RUNTIME_DIR=\$SCRATCH/.runtime&lt;br /&gt;
export JUPYTER_CONFIG_DIR=\$SCRATCH/.config/.jupyter&lt;br /&gt;
jupyter \${1:-notebook} --ip \$(hostname -f) --no-browser --notebook-dir=\$PWD&lt;br /&gt;
EOF&lt;br /&gt;
$ chmod +x  $HOME/.virtualenvs/jupteach/bin/notebook.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2. To run the jupyter server on a compute node, start an interactive session with the &amp;lt;tt&amp;gt;debugjob&amp;lt;/tt&amp;gt; command and then launch the jupyter server:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
$ debugjob -n 16                    # use less if you need less cores.&lt;br /&gt;
$ cd $SCRATCH                       # $HOME is read-only, so move to $SCRATCH&lt;br /&gt;
$ $HOME/.virtualenvs/jupteach/bin/notebook.sh  # add the argument &amp;quot;lab&amp;quot; to start with the jupyter lab&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
Make sure you note down (a) the name of the compute node that you got allocated (they start with &amp;quot;&amp;lt;tt&amp;gt;teach&amp;lt;/tt&amp;gt;&amp;quot; followed by a 2-digit number), and (b) the number following the compute nodes after the colon (usually this is 8888, but it can be another, higher number); this is the PORT  and (c) the last URL that the notebook.sh tells you to use to connect.&lt;br /&gt;
&lt;br /&gt;
4. To connect to this jupyter server running on a teach compute node, which is not accessible from the internet, in a different terminal on you own computer, you must reconnect to the Teach cluster with a port-forwarding&lt;br /&gt;
tunnel to the compute node on which jupyter is running:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
$ ssh -LPORT:teachXX:PORT -o ControlMaster=no USERNAME@teach.scinet.utoronto.ca  -N&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;tt&amp;gt;teachXX&amp;lt;/tt&amp;gt; is to be replaced by the name of the compute node (point (a) above), PORT is to be replaced by the port number that notebook.sh showed, and &amp;lt;tt&amp;gt;USERNAME&amp;lt;/t&amp;gt; should be your teach account username. This command should just &amp;quot;hang&amp;quot; there, it only serves to forward port number PORT (usually 8888) to port PORT (usually 8888) on the compute node.  &lt;br /&gt;
&lt;br /&gt;
Finally, point your browser to the URL that the &amp;lt;tt&amp;gt;notebook.sh&amp;lt;/tt&amp;gt; command printed out (point (b) above), i.e., the one with 127.0.0.1 in it.&lt;br /&gt;
--&amp;gt;&lt;/div&gt;</summary>
		<author><name>Rzon</name></author>
	</entry>
	<entry>
		<id>https://docs.scinet.utoronto.ca/index.php?title=Teach&amp;diff=7463</id>
		<title>Teach</title>
		<link rel="alternate" type="text/html" href="https://docs.scinet.utoronto.ca/index.php?title=Teach&amp;diff=7463"/>
		<updated>2026-01-05T14:29:40Z</updated>

		<summary type="html">&lt;p&gt;Rzon: /* Main changes from Teach's predecessor */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;This page describes the usage of the new Teach cluster, installed in Feb 2025.  It is currently still somewhat in beta phase.&lt;br /&gt;
&lt;br /&gt;
{{Infobox Computer&lt;br /&gt;
|image=[[Image:Ibm_idataplex_dx360_m4.jpg|center|300px|thumb]] &lt;br /&gt;
|name=Teach Cluster &lt;br /&gt;
|installed=(orig Mar 2020), Feb 2025&lt;br /&gt;
|operatingsystem= Linux (Rocky 9.6)&lt;br /&gt;
|loginnode=teach-login01&lt;br /&gt;
|nnodes=64 &lt;br /&gt;
|rampernode=188 GiB / 202 GB &lt;br /&gt;
|corespernode=40 &lt;br /&gt;
|interconnect=Infiniband (EDR)&lt;br /&gt;
|vendorcompilers=gcc,intel&lt;br /&gt;
|queuetype=slurm&lt;br /&gt;
}}&lt;br /&gt;
&lt;br /&gt;
== Teaching Cluster ==&lt;br /&gt;
&lt;br /&gt;
SciNet has assembled some older compute hardware into a smaller 2560-core cluster provided primarily for teaching purposes. It has currently no GPU capability.&lt;br /&gt;
&lt;br /&gt;
The Teach cluster is configured similarly to the coming SciNet production systems Trillium, however it uses hardware repurposed from its predecessor, [[Niagara_Quickstart|Niagara]].  This system should not be used for production work as the queuing policies are designed to provide fast job turnover and limit the amount of resources one person can use at a time.     Questions about its use or problems should be sent to '''support@scinet.utoronto.ca'''.&lt;br /&gt;
&lt;br /&gt;
This Teach cluster is setup differently from its predecessor.  See below for the main changes. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Specifications==&lt;br /&gt;
&lt;br /&gt;
This cluster currently consists of 64 repurposed x86_64 nodes each with 40 cores (from two 20 core Intel CascadeLake CPUs) running at 2.5GHz with 188GB of RAM per node. &lt;br /&gt;
The nodes are interconnected with 1:1 non-blocking EDR Infiniband for MPI communications, and disk I/O to a separate view of the VAST file system.  In total, this cluster contains 320 cores, but there are plans to expand it if demand warrants it.&lt;br /&gt;
&lt;br /&gt;
== Login/Devel Node ==&lt;br /&gt;
&lt;br /&gt;
Teach runs Rocky Linux 9.  You will need to be somewhat familiar with Linux systems to work on Teach.  If you are not, it will be worth your time to review our [https://education.scinet.utoronto.ca/tag/index.php?tag=SCMP101 Introduction to Linux Shell] class.&lt;br /&gt;
&lt;br /&gt;
As with all SciNet and {{Alliance}} systems, access to Teach is done via [[SSH]] (secure shell) only.  Open a terminal window (e.g. using [https://docs.alliancecan.ca/wiki/Connecting_with_PuTTY PuTTY] or  [https://docs.alliancecan.ca/wiki/Connecting_with_MobaXTerm MobaXTerm] on Windows), and type&lt;br /&gt;
 ssh -Y USERNAME@teach.scinet.utoronto.ca&lt;br /&gt;
This will bring directly to the command line of '''&amp;lt;tt&amp;gt;teach-login01&amp;lt;/tt&amp;gt;''' or '''&amp;lt;tt&amp;gt;teach-login02&amp;lt;/tt&amp;gt;''', which are the gateway/devel nodes for this cluster.  &lt;br /&gt;
On these nodes, you can compile, do short tests, and submit your jobs to the queue.&lt;br /&gt;
&lt;br /&gt;
The first time you login to Teach cluster, please make sure to check if the login node ssh key fingerprint&lt;br /&gt;
matches. [[Teach_fingerprints | See here how]].&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The login nodes are shared between students of a          &lt;br /&gt;
number of different courses. Use this node to develop and compile                  &lt;br /&gt;
code, to run short tests, and to submit computations to the scheduler (see below).&lt;br /&gt;
&lt;br /&gt;
Note that access to the teach cluster is restricted to temporary accounts that start with the prefix '''lcl_uot''' + the course code + '''s''', and a number.  Passwords for these accounts can be changed on the [https://portal.scinet.utoronto.ca/portaluserlogin SciNet user portal]. On the same site, users can upload a public ssh key if they want to connect using ssh keys.&lt;br /&gt;
&lt;br /&gt;
== Software Modules ==&lt;br /&gt;
&lt;br /&gt;
Other than essentials, all installed software is made available [[Using_modules | using module commands]]. These modules set environment variables (PATH, etc.), allowing multiple, conflicting versions of a given package to be available.  A detailed explanation of the module system can be [[Using_modules | found on the modules page]].  &lt;br /&gt;
&lt;br /&gt;
The Teach cluster makes the same [https://docs.alliancecan.ca/wiki/Available_software modules available] as on the [https://docs.alliancecan.ca General Purpose clusters of the Digital Research Alliance of Canada], with one caveat.  On Teach, by default, only the &amp;quot;gentoo&amp;quot; module is loaded, which provides basic OS-level functionality.  &lt;br /&gt;
&lt;br /&gt;
Common module subcommands are:&lt;br /&gt;
&lt;br /&gt;
* &amp;lt;code&amp;gt;module load &amp;lt;module-name&amp;gt;&amp;lt;/code&amp;gt;: load the default version of a particular software.&lt;br /&gt;
* &amp;lt;code&amp;gt;module load &amp;lt;module-name&amp;gt;/&amp;lt;module-version&amp;gt;&amp;lt;/code&amp;gt;: load a specific version of a particular software.&lt;br /&gt;
* &amp;lt;code&amp;gt;module purge&amp;lt;/code&amp;gt;: unload all currently loaded modules.&lt;br /&gt;
* &amp;lt;code&amp;gt;module spider&amp;lt;/code&amp;gt; (or &amp;lt;code&amp;gt;module spider &amp;lt;module-name&amp;gt;&amp;lt;/code&amp;gt;): list available software packages.&lt;br /&gt;
* &amp;lt;code&amp;gt;module avail&amp;lt;/code&amp;gt;: list loadable software packages.&lt;br /&gt;
* &amp;lt;code&amp;gt;module list&amp;lt;/code&amp;gt;: list loaded modules.&lt;br /&gt;
&lt;br /&gt;
For example, to make the GNU compilers (gcc, g++ and gfortran) available, you should type&lt;br /&gt;
&lt;br /&gt;
 module load gcc&lt;br /&gt;
&lt;br /&gt;
while the Intel compilers (icc, icpc and ifort) can be loaded by&lt;br /&gt;
&lt;br /&gt;
 module load intel&lt;br /&gt;
&lt;br /&gt;
To get the default modules that are loaded on the General Purpose clusters, you can load the &amp;quot;StdEnv&amp;quot; module. &lt;br /&gt;
&lt;br /&gt;
Along with modifying common environment variables, such as the PATH, these modules also create an '''EBROOT&amp;lt;MODULENAME&amp;gt;''' environment variable, which can be used to access commonly needed software directories, such as /include and /lib.&lt;br /&gt;
&lt;br /&gt;
There are handy abbreviations for the module commands. &amp;lt;code&amp;gt;ml&amp;lt;/code&amp;gt; is the same as &amp;lt;code&amp;gt;module list&amp;lt;/code&amp;gt;, and &amp;lt;code&amp;gt;ml &amp;lt;module-name&amp;gt;&amp;lt;/code&amp;gt; is the same as &amp;lt;code&amp;gt;module load &amp;lt;module-name&amp;gt;&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
A list of available software modules can be on [https://docs.alliancecan.ca/wiki/Available_software found on this page].&lt;br /&gt;
&lt;br /&gt;
There are a few addition modules available as well, and more can be made available upon demand of the course instructors.  Currently, the only additional modules are:&lt;br /&gt;
&lt;br /&gt;
 catch2/3.3.1      - A C++ test framework for unit-tests, TDD and BDD using C++14 and later.&lt;br /&gt;
 misopy/0.5.2      - A probabilistic framework to analysize RNA-Seq data.&lt;br /&gt;
 palemoon/33.6.0.1 - A web browser&lt;br /&gt;
&lt;br /&gt;
== Interactive jobs ==&lt;br /&gt;
&lt;br /&gt;
For a interactive sessions on a compute node of the teach cluster that give access to non-shared resources, use the 'debugjob' command. &lt;br /&gt;
 teach01:~$ debugjob -n C&lt;br /&gt;
where C is the number of cores. An interactive session defaults to four hours when using at most one node (C&amp;lt;=40), and becomes 60 minutes when using four nodes (i.e., 120&amp;lt;C&amp;lt;=160), which is the maximum number of nodes allowed for an interactive session by debugjob.&lt;br /&gt;
                                                             &lt;br /&gt;
For a short interactive sessions on a dedicated compute node of the teach cluster, use the 'debugjob' command as follows: &lt;br /&gt;
 teach01:~$ debugjob N&lt;br /&gt;
where N is the number of nodes.  On the Teach cluster, this is equivalent to &amp;lt;tt&amp;gt;debugjob -n 40*N &amp;lt;/tt&amp;gt;. The positive integer number &amp;lt;tt&amp;gt;N&amp;lt;/tt&amp;gt; can at most be 4.&lt;br /&gt;
&lt;br /&gt;
If no arguments are given to &amp;lt;tt&amp;gt;debugjob&amp;lt;/tt&amp;gt;, it allocates a single core on a Teach compute node.&lt;br /&gt;
&lt;br /&gt;
There are limits on the resources you can get with a debugjob, and how long you can get them.  No debugjob can run longer than four hours or use more than 160 cores, and each user can only run one at a time.  For longer computations, jobs must be submitted to the scheduler.&lt;br /&gt;
&lt;br /&gt;
== Submit a Job ==&lt;br /&gt;
&lt;br /&gt;
Teach uses SLURM as its job scheduler.  More-advanced details of how to interact with the scheduler can be found on the [[Slurm | Slurm page]].&lt;br /&gt;
&lt;br /&gt;
You submit jobs from a login node by passing a script to the sbatch command:&lt;br /&gt;
&lt;br /&gt;
 teach-login01:~$ sbatch jobscript.sh&lt;br /&gt;
&lt;br /&gt;
This puts the job in the queue. It will run on the compute nodes in due course.&lt;br /&gt;
&lt;br /&gt;
Note:&lt;br /&gt;
* Make sure to adjust accordingly the flags --ntasks-per-node or --ntasks together with --nodes for the examples found at [[Slurm | Slurm page]]. &lt;br /&gt;
* The current slurm configuration of the teach cluster allocates compute resources by core as opposed to by node. That means your tasks might land on nodes that have other jobs running, i.e. they might share the node. If you want to avoid that, make sure to add the following directive in your submitting script: #SBATCH --exclusive. This forces your job to use the compute nodes exclusively.&lt;br /&gt;
* The maximum wall time is currently set to 4 hours.&lt;br /&gt;
* There are 2 queues available: Compute queue and debug queue. Their usage limits are listed on the table below.&lt;br /&gt;
&lt;br /&gt;
== Limits ==&lt;br /&gt;
There are limits to the size and duration of your jobs, the number of jobs you can run and the number of jobs you can have queued. It also matters in which 'partition' the jobs runs. 'Partitions' are SLURM-speak for use cases. You specify the partition with the -p parameter to sbatch or salloc, but if you do not specify one, your job will run in the compute partition, which is the most common case.&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!Usage&lt;br /&gt;
!Partition&lt;br /&gt;
!Running jobs&lt;br /&gt;
!Submitted jobs (incl. running)&lt;br /&gt;
!Min. size of jobs&lt;br /&gt;
!Max. size of jobs&lt;br /&gt;
!Min. walltime&lt;br /&gt;
!Max. walltime &lt;br /&gt;
|-&lt;br /&gt;
|Interactive testing or troubleshooting || debug || 1 || 1 || 1 core || 4 nodes (160 cores)|| N/A || 4 hours&lt;br /&gt;
|-&lt;br /&gt;
|Compute jobs ||compute || 1 || 12 || 1 core || 4 nodes (160 cores)|| 15 minutes || 4 hours&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
Within these limits, jobs may still have to wait in the queue.  Although there are no allocations on the teach cluster, the waiting time still depends on many factors, such as the number of nodes and the wall time, how many other jobs are waiting in the queue, and whether a job can fill an otherwise unused spot in the schedule.&lt;br /&gt;
&lt;br /&gt;
== Main changes from Teach's predecessor ==&lt;br /&gt;
&lt;br /&gt;
This section is intended for instructors that may have use SciNet's previous Teach cluster.&lt;br /&gt;
&lt;br /&gt;
Although the cluster is once again called ''Teach'' and you connect with to teach.scinet.utoronto.ca as before, the system is setup differently from the previous Teach cluster in the following ways:&lt;br /&gt;
&lt;br /&gt;
* There are now 2 dedicated login nodes, teach-login01 and teach-login02.  &lt;br /&gt;
* The ssh fingerprints for these login nodes can be found on [[Teach_fingerprints]].&lt;br /&gt;
* The compute nodes have 40 cores.  As before, you can request jobs by number of cores.&lt;br /&gt;
* Only temporary lcl_uot.... accounts can log in.&lt;br /&gt;
* Only the home directories of those account are mounted. &lt;br /&gt;
* In particular, the file systems from the other SciNet compute clusters (Niagara, Mist, Trillium,... ) are not and will not be mounted.  You'll need to copy over any files that you need to use on the Teach cluster.&lt;br /&gt;
* There is no $SCRATCH.  You can do all your work on $HOME, which is writable from compute nodes&lt;br /&gt;
* The software stack is the one supplied by the Alliance.  There is no need to load 'CCEnv' to get them.&lt;br /&gt;
* But as before, if you're missing a module, we can still install it for you.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- Do this later&lt;br /&gt;
== Running Jupyter on a Teach Compute Node ==  &lt;br /&gt;
&lt;br /&gt;
1. To be able to run Jupyter on a compute node, you must first (a) install it inside a virtual environment, (b) enable a way for jupyter to seemingly write to a specific directory on $HOME, and (c) create a little helper script called &amp;lt;tt&amp;gt;notebook.sh&amp;lt;/tt&amp;gt; that will be used to start the jupyter server in step 2.  These are the commands that you should use for the installation (which you should do only once, on the Teach login node):&lt;br /&gt;
&lt;br /&gt;
(a) Create virtual env&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
$ module load python/3.9.10&lt;br /&gt;
$ virtualenv --system-site-packages $HOME/.virtualenvs/jupteach&lt;br /&gt;
$ source $HOME/.virtualenvs/jupteach/bin/activate&lt;br /&gt;
$ pip install jupyter jupyterlab&lt;br /&gt;
$ deactivate&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
You can choose another directory than &amp;lt;tt&amp;gt;$HOME/.virtualenvs/jupteach&amp;lt;/tt&amp;gt; for where to create the virtual environment, but you need to be consistent and use the same directory everywhere below.&lt;br /&gt;
&lt;br /&gt;
(b) Make a writable 'runtime' directory for Jupyter. &lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
$ mkdir -p $HOME/.local/share/jupyter/runtime &lt;br /&gt;
$ mv -f $HOME/.local/share/jupyter/runtime $SCRATCH/jupyter_runtime || mkdir $SCRATCH/jupyter_runtime&lt;br /&gt;
$ ln -sT $SCRATCH/jupyter_runtime $HOME/.local/share/jupyter/runtime&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
(c) Create a launch script to use on the compute nodes:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
$ cat &amp;gt; $HOME/.virtualenvs/jupteach/bin/notebook.sh &amp;lt;&amp;lt;EOF&lt;br /&gt;
#!/bin/bash&lt;br /&gt;
source \$HOME/.virtualenvs/jupteach/bin/activate&lt;br /&gt;
export XDG_DATA_HOME=\$SCRATCH/.share&lt;br /&gt;
export XDG_CACHE_HOME=\$SCRATCH/.cache&lt;br /&gt;
export XDG_CONFIG_HOME=\$SCRATCH/.config&lt;br /&gt;
export XDG_RUNTIME_DIR=\$SCRATCH/.runtime&lt;br /&gt;
export JUPYTER_CONFIG_DIR=\$SCRATCH/.config/.jupyter&lt;br /&gt;
jupyter \${1:-notebook} --ip \$(hostname -f) --no-browser --notebook-dir=\$PWD&lt;br /&gt;
EOF&lt;br /&gt;
$ chmod +x  $HOME/.virtualenvs/jupteach/bin/notebook.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2. To run the jupyter server on a compute node, start an interactive session with the &amp;lt;tt&amp;gt;debugjob&amp;lt;/tt&amp;gt; command and then launch the jupyter server:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
$ debugjob -n 16                    # use less if you need less cores.&lt;br /&gt;
$ cd $SCRATCH                       # $HOME is read-only, so move to $SCRATCH&lt;br /&gt;
$ $HOME/.virtualenvs/jupteach/bin/notebook.sh  # add the argument &amp;quot;lab&amp;quot; to start with the jupyter lab&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
Make sure you note down (a) the name of the compute node that you got allocated (they start with &amp;quot;&amp;lt;tt&amp;gt;teach&amp;lt;/tt&amp;gt;&amp;quot; followed by a 2-digit number), and (b) the number following the compute nodes after the colon (usually this is 8888, but it can be another, higher number); this is the PORT  and (c) the last URL that the notebook.sh tells you to use to connect.&lt;br /&gt;
&lt;br /&gt;
4. To connect to this jupyter server running on a teach compute node, which is not accessible from the internet, in a different terminal on you own computer, you must reconnect to the Teach cluster with a port-forwarding&lt;br /&gt;
tunnel to the compute node on which jupyter is running:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
$ ssh -LPORT:teachXX:PORT -o ControlMaster=no USERNAME@teach.scinet.utoronto.ca  -N&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;tt&amp;gt;teachXX&amp;lt;/tt&amp;gt; is to be replaced by the name of the compute node (point (a) above), PORT is to be replaced by the port number that notebook.sh showed, and &amp;lt;tt&amp;gt;USERNAME&amp;lt;/t&amp;gt; should be your teach account username. This command should just &amp;quot;hang&amp;quot; there, it only serves to forward port number PORT (usually 8888) to port PORT (usually 8888) on the compute node.  &lt;br /&gt;
&lt;br /&gt;
Finally, point your browser to the URL that the &amp;lt;tt&amp;gt;notebook.sh&amp;lt;/tt&amp;gt; command printed out (point (b) above), i.e., the one with 127.0.0.1 in it.&lt;br /&gt;
--&amp;gt;&lt;/div&gt;</summary>
		<author><name>Rzon</name></author>
	</entry>
	<entry>
		<id>https://docs.scinet.utoronto.ca/index.php?title=Teach&amp;diff=7403</id>
		<title>Teach</title>
		<link rel="alternate" type="text/html" href="https://docs.scinet.utoronto.ca/index.php?title=Teach&amp;diff=7403"/>
		<updated>2025-12-11T19:58:15Z</updated>

		<summary type="html">&lt;p&gt;Rzon: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;This page describes the usage of the new Teach cluster, installed in Feb 2025.  It is currently still somewhat in beta phase.&lt;br /&gt;
&lt;br /&gt;
{{Infobox Computer&lt;br /&gt;
|image=[[Image:Ibm_idataplex_dx360_m4.jpg|center|300px|thumb]] &lt;br /&gt;
|name=Teach Cluster &lt;br /&gt;
|installed=(orig Mar 2020), Feb 2025&lt;br /&gt;
|operatingsystem= Linux (Rocky 9.6)&lt;br /&gt;
|loginnode=teach-login01&lt;br /&gt;
|nnodes=64 &lt;br /&gt;
|rampernode=188 GiB / 202 GB &lt;br /&gt;
|corespernode=40 &lt;br /&gt;
|interconnect=Infiniband (EDR)&lt;br /&gt;
|vendorcompilers=gcc,intel&lt;br /&gt;
|queuetype=slurm&lt;br /&gt;
}}&lt;br /&gt;
&lt;br /&gt;
== Teaching Cluster ==&lt;br /&gt;
&lt;br /&gt;
SciNet has assembled some older compute hardware into a smaller 2560-core cluster provided primarily for teaching purposes. It has currently no GPU capability.&lt;br /&gt;
&lt;br /&gt;
The Teach cluster is configured similarly to the coming SciNet production systems Trillium, however it uses hardware repurposed from its predecessor, [[Niagara_Quickstart|Niagara]].  This system should not be used for production work as the queuing policies are designed to provide fast job turnover and limit the amount of resources one person can use at a time.     Questions about its use or problems should be sent to '''support@scinet.utoronto.ca'''.&lt;br /&gt;
&lt;br /&gt;
This Teach cluster is setup differently from its predecessor.  See below for the main changes. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Specifications==&lt;br /&gt;
&lt;br /&gt;
This cluster currently consists of 64 repurposed x86_64 nodes each with 40 cores (from two 20 core Intel CascadeLake CPUs) running at 2.5GHz with 188GB of RAM per node. &lt;br /&gt;
The nodes are interconnected with 1:1 non-blocking EDR Infiniband for MPI communications, and disk I/O to a separate view of the VAST file system.  In total, this cluster contains 320 cores, but there are plans to expand it if demand warrants it.&lt;br /&gt;
&lt;br /&gt;
== Login/Devel Node ==&lt;br /&gt;
&lt;br /&gt;
Teach runs Rocky Linux 9.  You will need to be somewhat familiar with Linux systems to work on Teach.  If you are not, it will be worth your time to review our [https://education.scinet.utoronto.ca/tag/index.php?tag=SCMP101 Introduction to Linux Shell] class.&lt;br /&gt;
&lt;br /&gt;
As with all SciNet and {{Alliance}} systems, access to Teach is done via [[SSH]] (secure shell) only.  Open a terminal window (e.g. using [https://docs.alliancecan.ca/wiki/Connecting_with_PuTTY PuTTY] or  [https://docs.alliancecan.ca/wiki/Connecting_with_MobaXTerm MobaXTerm] on Windows), and type&lt;br /&gt;
 ssh -Y USERNAME@teach.scinet.utoronto.ca&lt;br /&gt;
This will bring directly to the command line of '''&amp;lt;tt&amp;gt;teach-login01&amp;lt;/tt&amp;gt;''' or '''&amp;lt;tt&amp;gt;teach-login02&amp;lt;/tt&amp;gt;''', which are the gateway/devel nodes for this cluster.  &lt;br /&gt;
On these nodes, you can compile, do short tests, and submit your jobs to the queue.&lt;br /&gt;
&lt;br /&gt;
The first time you login to Teach cluster, please make sure to check if the login node ssh key fingerprint&lt;br /&gt;
matches. [[Teach_fingerprints | See here how]].&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The login nodes are shared between students of a          &lt;br /&gt;
number of different courses. Use this node to develop and compile                  &lt;br /&gt;
code, to run short tests, and to submit computations to the scheduler (see below).&lt;br /&gt;
&lt;br /&gt;
Note that access to the teach cluster is restricted to temporary accounts that start with the prefix '''lcl_uot''' + the course code + '''s''', and a number.  Passwords for these accounts can be changed on the [https://portal.scinet.utoronto.ca/portaluserlogin SciNet user portal]. On the same site, users can upload a public ssh key if they want to connect using ssh keys.&lt;br /&gt;
&lt;br /&gt;
== Software Modules ==&lt;br /&gt;
&lt;br /&gt;
Other than essentials, all installed software is made available [[Using_modules | using module commands]]. These modules set environment variables (PATH, etc.), allowing multiple, conflicting versions of a given package to be available.  A detailed explanation of the module system can be [[Using_modules | found on the modules page]].  &lt;br /&gt;
&lt;br /&gt;
The Teach cluster makes the same [https://docs.alliancecan.ca/wiki/Available_software modules available] as on the [https://docs.alliancecan.ca General Purpose clusters of the Digital Research Alliance of Canada], with one caveat.  On Teach, by default, only the &amp;quot;gentoo&amp;quot; module is loaded, which provides basic OS-level functionality.  &lt;br /&gt;
&lt;br /&gt;
Common module subcommands are:&lt;br /&gt;
&lt;br /&gt;
* &amp;lt;code&amp;gt;module load &amp;lt;module-name&amp;gt;&amp;lt;/code&amp;gt;: load the default version of a particular software.&lt;br /&gt;
* &amp;lt;code&amp;gt;module load &amp;lt;module-name&amp;gt;/&amp;lt;module-version&amp;gt;&amp;lt;/code&amp;gt;: load a specific version of a particular software.&lt;br /&gt;
* &amp;lt;code&amp;gt;module purge&amp;lt;/code&amp;gt;: unload all currently loaded modules.&lt;br /&gt;
* &amp;lt;code&amp;gt;module spider&amp;lt;/code&amp;gt; (or &amp;lt;code&amp;gt;module spider &amp;lt;module-name&amp;gt;&amp;lt;/code&amp;gt;): list available software packages.&lt;br /&gt;
* &amp;lt;code&amp;gt;module avail&amp;lt;/code&amp;gt;: list loadable software packages.&lt;br /&gt;
* &amp;lt;code&amp;gt;module list&amp;lt;/code&amp;gt;: list loaded modules.&lt;br /&gt;
&lt;br /&gt;
For example, to make the GNU compilers (gcc, g++ and gfortran) available, you should type&lt;br /&gt;
&lt;br /&gt;
 module load gcc&lt;br /&gt;
&lt;br /&gt;
while the Intel compilers (icc, icpc and ifort) can be loaded by&lt;br /&gt;
&lt;br /&gt;
 module load intel&lt;br /&gt;
&lt;br /&gt;
To get the default modules that are loaded on the General Purpose clusters, you can load the &amp;quot;StdEnv&amp;quot; module. &lt;br /&gt;
&lt;br /&gt;
Along with modifying common environment variables, such as the PATH, these modules also create an '''EBROOT&amp;lt;MODULENAME&amp;gt;''' environment variable, which can be used to access commonly needed software directories, such as /include and /lib.&lt;br /&gt;
&lt;br /&gt;
There are handy abbreviations for the module commands. &amp;lt;code&amp;gt;ml&amp;lt;/code&amp;gt; is the same as &amp;lt;code&amp;gt;module list&amp;lt;/code&amp;gt;, and &amp;lt;code&amp;gt;ml &amp;lt;module-name&amp;gt;&amp;lt;/code&amp;gt; is the same as &amp;lt;code&amp;gt;module load &amp;lt;module-name&amp;gt;&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
A list of available software modules can be on [https://docs.alliancecan.ca/wiki/Available_software found on this page].&lt;br /&gt;
&lt;br /&gt;
There are a few addition modules available as well, and more can be made available upon demand of the course instructors.  Currently, the only additional modules are:&lt;br /&gt;
&lt;br /&gt;
 catch2/3.3.1      - A C++ test framework for unit-tests, TDD and BDD using C++14 and later.&lt;br /&gt;
 misopy/0.5.2      - A probabilistic framework to analysize RNA-Seq data.&lt;br /&gt;
 palemoon/33.6.0.1 - A web browser&lt;br /&gt;
&lt;br /&gt;
== Interactive jobs ==&lt;br /&gt;
&lt;br /&gt;
For a interactive sessions on a compute node of the teach cluster that give access to non-shared resources, use the 'debugjob' command. &lt;br /&gt;
 teach01:~$ debugjob -n C&lt;br /&gt;
where C is the number of cores. An interactive session defaults to four hours when using at most one node (C&amp;lt;=40), and becomes 60 minutes when using four nodes (i.e., 120&amp;lt;C&amp;lt;=160), which is the maximum number of nodes allowed for an interactive session by debugjob.&lt;br /&gt;
                                                             &lt;br /&gt;
For a short interactive sessions on a dedicated compute node of the teach cluster, use the 'debugjob' command as follows: &lt;br /&gt;
 teach01:~$ debugjob N&lt;br /&gt;
where N is the number of nodes.  On the Teach cluster, this is equivalent to &amp;lt;tt&amp;gt;debugjob -n 40*N &amp;lt;/tt&amp;gt;. The positive integer number &amp;lt;tt&amp;gt;N&amp;lt;/tt&amp;gt; can at most be 4.&lt;br /&gt;
&lt;br /&gt;
If no arguments are given to &amp;lt;tt&amp;gt;debugjob&amp;lt;/tt&amp;gt;, it allocates a single core on a Teach compute node.&lt;br /&gt;
&lt;br /&gt;
There are limits on the resources you can get with a debugjob, and how long you can get them.  No debugjob can run longer than four hours or use more than 160 cores, and each user can only run one at a time.  For longer computations, jobs must be submitted to the scheduler.&lt;br /&gt;
&lt;br /&gt;
== Submit a Job ==&lt;br /&gt;
&lt;br /&gt;
Teach uses SLURM as its job scheduler.  More-advanced details of how to interact with the scheduler can be found on the [[Slurm | Slurm page]].&lt;br /&gt;
&lt;br /&gt;
You submit jobs from a login node by passing a script to the sbatch command:&lt;br /&gt;
&lt;br /&gt;
 teach-login01:~$ sbatch jobscript.sh&lt;br /&gt;
&lt;br /&gt;
This puts the job in the queue. It will run on the compute nodes in due course.&lt;br /&gt;
&lt;br /&gt;
Note:&lt;br /&gt;
* Make sure to adjust accordingly the flags --ntasks-per-node or --ntasks together with --nodes for the examples found at [[Slurm | Slurm page]]. &lt;br /&gt;
* The current slurm configuration of the teach cluster allocates compute resources by core as opposed to by node. That means your tasks might land on nodes that have other jobs running, i.e. they might share the node. If you want to avoid that, make sure to add the following directive in your submitting script: #SBATCH --exclusive. This forces your job to use the compute nodes exclusively.&lt;br /&gt;
* The maximum wall time is currently set to 4 hours.&lt;br /&gt;
* There are 2 queues available: Compute queue and debug queue. Their usage limits are listed on the table below.&lt;br /&gt;
&lt;br /&gt;
== Limits ==&lt;br /&gt;
There are limits to the size and duration of your jobs, the number of jobs you can run and the number of jobs you can have queued. It also matters in which 'partition' the jobs runs. 'Partitions' are SLURM-speak for use cases. You specify the partition with the -p parameter to sbatch or salloc, but if you do not specify one, your job will run in the compute partition, which is the most common case.&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!Usage&lt;br /&gt;
!Partition&lt;br /&gt;
!Running jobs&lt;br /&gt;
!Submitted jobs (incl. running)&lt;br /&gt;
!Min. size of jobs&lt;br /&gt;
!Max. size of jobs&lt;br /&gt;
!Min. walltime&lt;br /&gt;
!Max. walltime &lt;br /&gt;
|-&lt;br /&gt;
|Interactive testing or troubleshooting || debug || 1 || 1 || 1 core || 4 nodes (160 cores)|| N/A || 4 hours&lt;br /&gt;
|-&lt;br /&gt;
|Compute jobs ||compute || 1 || 12 || 1 core || 4 nodes (160 cores)|| 15 minutes || 4 hours&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
Within these limits, jobs may still have to wait in the queue.  Although there are no allocations on the teach cluster, the waiting time still depends on many factors, such as the number of nodes and the wall time, how many other jobs are waiting in the queue, and whether a job can fill an otherwise unused spot in the schedule.&lt;br /&gt;
&lt;br /&gt;
== Main changes from Teach's predecessor ==&lt;br /&gt;
&lt;br /&gt;
Although the cluster is once again called ''Teach'' and you connect with to teach.scinet.utoronto.ca as before, the system is setup differently from the previous Teach cluster in the following ways:&lt;br /&gt;
&lt;br /&gt;
* There are now 2 dedicated login nodes, teach-login01 and teach-login02.  &lt;br /&gt;
* The ssh fingerprints for these login nodes can be found on [[Teach_fingerprints]].&lt;br /&gt;
* The compute nodes have 40 cores.  As before, you can request jobs by number of cores.&lt;br /&gt;
* Only temporary lcl_uot.... accounts can log in.&lt;br /&gt;
* Only the home directories of those account are mounted. &lt;br /&gt;
* In particular, the file systems from the other SciNet compute clusters (Niagara, Mist, Trillium,... ) are not and will not be mounted.  You'll need to copy over any files that you need to use on the Teach cluster.&lt;br /&gt;
* There is no $SCRATCH.  You can do all your work on $HOME, which is writable from compute nodes&lt;br /&gt;
* The software stack is the one supplied by the Alliance.  There is no need to load 'CCEnv' to get them.&lt;br /&gt;
* But as before, if you're missing a module, we can still install it for you.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- Do this later&lt;br /&gt;
== Running Jupyter on a Teach Compute Node ==  &lt;br /&gt;
&lt;br /&gt;
1. To be able to run Jupyter on a compute node, you must first (a) install it inside a virtual environment, (b) enable a way for jupyter to seemingly write to a specific directory on $HOME, and (c) create a little helper script called &amp;lt;tt&amp;gt;notebook.sh&amp;lt;/tt&amp;gt; that will be used to start the jupyter server in step 2.  These are the commands that you should use for the installation (which you should do only once, on the Teach login node):&lt;br /&gt;
&lt;br /&gt;
(a) Create virtual env&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
$ module load python/3.9.10&lt;br /&gt;
$ virtualenv --system-site-packages $HOME/.virtualenvs/jupteach&lt;br /&gt;
$ source $HOME/.virtualenvs/jupteach/bin/activate&lt;br /&gt;
$ pip install jupyter jupyterlab&lt;br /&gt;
$ deactivate&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
You can choose another directory than &amp;lt;tt&amp;gt;$HOME/.virtualenvs/jupteach&amp;lt;/tt&amp;gt; for where to create the virtual environment, but you need to be consistent and use the same directory everywhere below.&lt;br /&gt;
&lt;br /&gt;
(b) Make a writable 'runtime' directory for Jupyter. &lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
$ mkdir -p $HOME/.local/share/jupyter/runtime &lt;br /&gt;
$ mv -f $HOME/.local/share/jupyter/runtime $SCRATCH/jupyter_runtime || mkdir $SCRATCH/jupyter_runtime&lt;br /&gt;
$ ln -sT $SCRATCH/jupyter_runtime $HOME/.local/share/jupyter/runtime&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
(c) Create a launch script to use on the compute nodes:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
$ cat &amp;gt; $HOME/.virtualenvs/jupteach/bin/notebook.sh &amp;lt;&amp;lt;EOF&lt;br /&gt;
#!/bin/bash&lt;br /&gt;
source \$HOME/.virtualenvs/jupteach/bin/activate&lt;br /&gt;
export XDG_DATA_HOME=\$SCRATCH/.share&lt;br /&gt;
export XDG_CACHE_HOME=\$SCRATCH/.cache&lt;br /&gt;
export XDG_CONFIG_HOME=\$SCRATCH/.config&lt;br /&gt;
export XDG_RUNTIME_DIR=\$SCRATCH/.runtime&lt;br /&gt;
export JUPYTER_CONFIG_DIR=\$SCRATCH/.config/.jupyter&lt;br /&gt;
jupyter \${1:-notebook} --ip \$(hostname -f) --no-browser --notebook-dir=\$PWD&lt;br /&gt;
EOF&lt;br /&gt;
$ chmod +x  $HOME/.virtualenvs/jupteach/bin/notebook.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2. To run the jupyter server on a compute node, start an interactive session with the &amp;lt;tt&amp;gt;debugjob&amp;lt;/tt&amp;gt; command and then launch the jupyter server:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
$ debugjob -n 16                    # use less if you need less cores.&lt;br /&gt;
$ cd $SCRATCH                       # $HOME is read-only, so move to $SCRATCH&lt;br /&gt;
$ $HOME/.virtualenvs/jupteach/bin/notebook.sh  # add the argument &amp;quot;lab&amp;quot; to start with the jupyter lab&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
Make sure you note down (a) the name of the compute node that you got allocated (they start with &amp;quot;&amp;lt;tt&amp;gt;teach&amp;lt;/tt&amp;gt;&amp;quot; followed by a 2-digit number), and (b) the number following the compute nodes after the colon (usually this is 8888, but it can be another, higher number); this is the PORT  and (c) the last URL that the notebook.sh tells you to use to connect.&lt;br /&gt;
&lt;br /&gt;
4. To connect to this jupyter server running on a teach compute node, which is not accessible from the internet, in a different terminal on you own computer, you must reconnect to the Teach cluster with a port-forwarding&lt;br /&gt;
tunnel to the compute node on which jupyter is running:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
$ ssh -LPORT:teachXX:PORT -o ControlMaster=no USERNAME@teach.scinet.utoronto.ca  -N&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;tt&amp;gt;teachXX&amp;lt;/tt&amp;gt; is to be replaced by the name of the compute node (point (a) above), PORT is to be replaced by the port number that notebook.sh showed, and &amp;lt;tt&amp;gt;USERNAME&amp;lt;/t&amp;gt; should be your teach account username. This command should just &amp;quot;hang&amp;quot; there, it only serves to forward port number PORT (usually 8888) to port PORT (usually 8888) on the compute node.  &lt;br /&gt;
&lt;br /&gt;
Finally, point your browser to the URL that the &amp;lt;tt&amp;gt;notebook.sh&amp;lt;/tt&amp;gt; command printed out (point (b) above), i.e., the one with 127.0.0.1 in it.&lt;br /&gt;
--&amp;gt;&lt;/div&gt;</summary>
		<author><name>Rzon</name></author>
	</entry>
	<entry>
		<id>https://docs.scinet.utoronto.ca/index.php?title=Teach&amp;diff=7400</id>
		<title>Teach</title>
		<link rel="alternate" type="text/html" href="https://docs.scinet.utoronto.ca/index.php?title=Teach&amp;diff=7400"/>
		<updated>2025-12-11T19:56:17Z</updated>

		<summary type="html">&lt;p&gt;Rzon: /* Specifications */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;This page describes the usage of the new Teach cluster, installed in Feb 2025.  It is currently still somewhat in beta phase.&lt;br /&gt;
&lt;br /&gt;
{{Infobox Computer&lt;br /&gt;
|image=[[Image:Ibm_idataplex_dx360_m4.jpg|center|300px|thumb]] &lt;br /&gt;
|name=Teach Cluster &lt;br /&gt;
|installed=(orig Mar 2020), Feb 2025&lt;br /&gt;
|operatingsystem= Linux (Rocky 9.5)&lt;br /&gt;
|loginnode=teach-login01&lt;br /&gt;
|nnodes=8 &lt;br /&gt;
|rampernode=188 GiB / 202 GB &lt;br /&gt;
|corespernode=40 &lt;br /&gt;
|interconnect=Infiniband (EDR)&lt;br /&gt;
|vendorcompilers=gcc&lt;br /&gt;
|queuetype=slurm&lt;br /&gt;
}}&lt;br /&gt;
&lt;br /&gt;
== Teaching Cluster ==&lt;br /&gt;
&lt;br /&gt;
SciNet has assembled some older compute hardware into a small cluster provided primarily for teaching purposes. It is configured similarly to the coming SciNet production systems Trillium, however it uses hardware repurposed from its predecessor, [[Niagara_Quickstart|Niagara]].  This system should not be used for production work as the queuing policies are designed to provide fast job turnover and limit the amount of resources one person can use at a time.     Questions about its use or problems should be sent to '''support@scinet.utoronto.ca'''.&lt;br /&gt;
&lt;br /&gt;
This Teach cluster is setup differently from its predecessor.  See below for the main changes. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Specifications==&lt;br /&gt;
&lt;br /&gt;
This cluster currently consists of 64 repurposed x86_64 nodes each with 40 cores (from two 20 core Intel CascadeLake CPUs) running at 2.5GHz with 188GB of RAM per node. &lt;br /&gt;
The nodes are interconnected with 1:1 non-blocking EDR Infiniband for MPI communications, and disk I/O to a separate view of the VAST file system.  In total, this cluster contains 320 cores, but there are plans to expand it if demand warrants it.&lt;br /&gt;
&lt;br /&gt;
== Login/Devel Node ==&lt;br /&gt;
&lt;br /&gt;
Teach runs Rocky Linux 9.  You will need to be somewhat familiar with Linux systems to work on Teach.  If you are not, it will be worth your time to review our [https://education.scinet.utoronto.ca/tag/index.php?tag=SCMP101 Introduction to Linux Shell] class.&lt;br /&gt;
&lt;br /&gt;
As with all SciNet and {{Alliance}} systems, access to Teach is done via [[SSH]] (secure shell) only.  Open a terminal window (e.g. using [https://docs.alliancecan.ca/wiki/Connecting_with_PuTTY PuTTY] or  [https://docs.alliancecan.ca/wiki/Connecting_with_MobaXTerm MobaXTerm] on Windows), and type&lt;br /&gt;
 ssh -Y USERNAME@teach.scinet.utoronto.ca&lt;br /&gt;
This will bring directly to the command line of '''&amp;lt;tt&amp;gt;teach-login01&amp;lt;/tt&amp;gt;''' or '''&amp;lt;tt&amp;gt;teach-login02&amp;lt;/tt&amp;gt;''', which are the gateway/devel nodes for this cluster.  &lt;br /&gt;
On these nodes, you can compile, do short tests, and submit your jobs to the queue.&lt;br /&gt;
&lt;br /&gt;
The first time you login to Teach cluster, please make sure to check if the login node ssh key fingerprint&lt;br /&gt;
matches. [[Teach_fingerprints | See here how]].&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The login nodes are shared between students of a          &lt;br /&gt;
number of different courses. Use this node to develop and compile                  &lt;br /&gt;
code, to run short tests, and to submit computations to the scheduler (see below).&lt;br /&gt;
&lt;br /&gt;
Note that access to the teach cluster is restricted to temporary accounts that start with the prefix '''lcl_uot''' + the course code + '''s''', and a number.  Passwords for these accounts can be changed on the [https://portal.scinet.utoronto.ca/portaluserlogin SciNet user portal]. On the same site, users can upload a public ssh key if they want to connect using ssh keys.&lt;br /&gt;
&lt;br /&gt;
== Software Modules ==&lt;br /&gt;
&lt;br /&gt;
Other than essentials, all installed software is made available [[Using_modules | using module commands]]. These modules set environment variables (PATH, etc.), allowing multiple, conflicting versions of a given package to be available.  A detailed explanation of the module system can be [[Using_modules | found on the modules page]].  &lt;br /&gt;
&lt;br /&gt;
The Teach cluster makes the same [https://docs.alliancecan.ca/wiki/Available_software modules available] as on the [https://docs.alliancecan.ca General Purpose clusters of the Digital Research Alliance of Canada], with one caveat.  On Teach, by default, only the &amp;quot;gentoo&amp;quot; module is loaded, which provides basic OS-level functionality.  &lt;br /&gt;
&lt;br /&gt;
Common module subcommands are:&lt;br /&gt;
&lt;br /&gt;
* &amp;lt;code&amp;gt;module load &amp;lt;module-name&amp;gt;&amp;lt;/code&amp;gt;: load the default version of a particular software.&lt;br /&gt;
* &amp;lt;code&amp;gt;module load &amp;lt;module-name&amp;gt;/&amp;lt;module-version&amp;gt;&amp;lt;/code&amp;gt;: load a specific version of a particular software.&lt;br /&gt;
* &amp;lt;code&amp;gt;module purge&amp;lt;/code&amp;gt;: unload all currently loaded modules.&lt;br /&gt;
* &amp;lt;code&amp;gt;module spider&amp;lt;/code&amp;gt; (or &amp;lt;code&amp;gt;module spider &amp;lt;module-name&amp;gt;&amp;lt;/code&amp;gt;): list available software packages.&lt;br /&gt;
* &amp;lt;code&amp;gt;module avail&amp;lt;/code&amp;gt;: list loadable software packages.&lt;br /&gt;
* &amp;lt;code&amp;gt;module list&amp;lt;/code&amp;gt;: list loaded modules.&lt;br /&gt;
&lt;br /&gt;
For example, to make the GNU compilers (gcc, g++ and gfortran) available, you should type&lt;br /&gt;
&lt;br /&gt;
 module load gcc&lt;br /&gt;
&lt;br /&gt;
while the Intel compilers (icc, icpc and ifort) can be loaded by&lt;br /&gt;
&lt;br /&gt;
 module load intel&lt;br /&gt;
&lt;br /&gt;
To get the default modules that are loaded on the General Purpose clusters, you can load the &amp;quot;StdEnv&amp;quot; module. &lt;br /&gt;
&lt;br /&gt;
Along with modifying common environment variables, such as the PATH, these modules also create an '''EBROOT&amp;lt;MODULENAME&amp;gt;''' environment variable, which can be used to access commonly needed software directories, such as /include and /lib.&lt;br /&gt;
&lt;br /&gt;
There are handy abbreviations for the module commands. &amp;lt;code&amp;gt;ml&amp;lt;/code&amp;gt; is the same as &amp;lt;code&amp;gt;module list&amp;lt;/code&amp;gt;, and &amp;lt;code&amp;gt;ml &amp;lt;module-name&amp;gt;&amp;lt;/code&amp;gt; is the same as &amp;lt;code&amp;gt;module load &amp;lt;module-name&amp;gt;&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
A list of available software modules can be on [https://docs.alliancecan.ca/wiki/Available_software found on this page].&lt;br /&gt;
&lt;br /&gt;
There are a few addition modules available as well, and more can be made available upon demand of the course instructors.  Currently, the only additional modules are:&lt;br /&gt;
&lt;br /&gt;
 catch2/3.3.1      - A C++ test framework for unit-tests, TDD and BDD using C++14 and later.&lt;br /&gt;
 misopy/0.5.2      - A probabilistic framework to analysize RNA-Seq data.&lt;br /&gt;
 palemoon/33.6.0.1 - A web browser&lt;br /&gt;
&lt;br /&gt;
== Interactive jobs ==&lt;br /&gt;
&lt;br /&gt;
For a interactive sessions on a compute node of the teach cluster that give access to non-shared resources, use the 'debugjob' command. &lt;br /&gt;
 teach01:~$ debugjob -n C&lt;br /&gt;
where C is the number of cores. An interactive session defaults to four hours when using at most one node (C&amp;lt;=40), and becomes 60 minutes when using four nodes (i.e., 120&amp;lt;C&amp;lt;=160), which is the maximum number of nodes allowed for an interactive session by debugjob.&lt;br /&gt;
                                                             &lt;br /&gt;
For a short interactive sessions on a dedicated compute node of the teach cluster, use the 'debugjob' command as follows: &lt;br /&gt;
 teach01:~$ debugjob N&lt;br /&gt;
where N is the number of nodes.  On the Teach cluster, this is equivalent to &amp;lt;tt&amp;gt;debugjob -n 40*N &amp;lt;/tt&amp;gt;. The positive integer number &amp;lt;tt&amp;gt;N&amp;lt;/tt&amp;gt; can at most be 4.&lt;br /&gt;
&lt;br /&gt;
If no arguments are given to &amp;lt;tt&amp;gt;debugjob&amp;lt;/tt&amp;gt;, it allocates a single core on a Teach compute node.&lt;br /&gt;
&lt;br /&gt;
There are limits on the resources you can get with a debugjob, and how long you can get them.  No debugjob can run longer than four hours or use more than 160 cores, and each user can only run one at a time.  For longer computations, jobs must be submitted to the scheduler.&lt;br /&gt;
&lt;br /&gt;
== Submit a Job ==&lt;br /&gt;
&lt;br /&gt;
Teach uses SLURM as its job scheduler.  More-advanced details of how to interact with the scheduler can be found on the [[Slurm | Slurm page]].&lt;br /&gt;
&lt;br /&gt;
You submit jobs from a login node by passing a script to the sbatch command:&lt;br /&gt;
&lt;br /&gt;
 teach-login01:~$ sbatch jobscript.sh&lt;br /&gt;
&lt;br /&gt;
This puts the job in the queue. It will run on the compute nodes in due course.&lt;br /&gt;
&lt;br /&gt;
Note:&lt;br /&gt;
* Make sure to adjust accordingly the flags --ntasks-per-node or --ntasks together with --nodes for the examples found at [[Slurm | Slurm page]]. &lt;br /&gt;
* The current slurm configuration of the teach cluster allocates compute resources by core as opposed to by node. That means your tasks might land on nodes that have other jobs running, i.e. they might share the node. If you want to avoid that, make sure to add the following directive in your submitting script: #SBATCH --exclusive. This forces your job to use the compute nodes exclusively.&lt;br /&gt;
* The maximum wall time is currently set to 4 hours.&lt;br /&gt;
* There are 2 queues available: Compute queue and debug queue. Their usage limits are listed on the table below.&lt;br /&gt;
&lt;br /&gt;
== Limits ==&lt;br /&gt;
There are limits to the size and duration of your jobs, the number of jobs you can run and the number of jobs you can have queued. It also matters in which 'partition' the jobs runs. 'Partitions' are SLURM-speak for use cases. You specify the partition with the -p parameter to sbatch or salloc, but if you do not specify one, your job will run in the compute partition, which is the most common case.&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!Usage&lt;br /&gt;
!Partition&lt;br /&gt;
!Running jobs&lt;br /&gt;
!Submitted jobs (incl. running)&lt;br /&gt;
!Min. size of jobs&lt;br /&gt;
!Max. size of jobs&lt;br /&gt;
!Min. walltime&lt;br /&gt;
!Max. walltime &lt;br /&gt;
|-&lt;br /&gt;
|Interactive testing or troubleshooting || debug || 1 || 1 || 1 core || 4 nodes (160 cores)|| N/A || 4 hours&lt;br /&gt;
|-&lt;br /&gt;
|Compute jobs ||compute || 1 || 12 || 1 core || 4 nodes (160 cores)|| 15 minutes || 4 hours&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
Within these limits, jobs may still have to wait in the queue.  Although there are no allocations on the teach cluster, the waiting time still depends on many factors, such as the number of nodes and the wall time, how many other jobs are waiting in the queue, and whether a job can fill an otherwise unused spot in the schedule.&lt;br /&gt;
&lt;br /&gt;
== Main changes from Teach's predecessor ==&lt;br /&gt;
&lt;br /&gt;
Although the cluster is once again called ''Teach'' and you connect with to teach.scinet.utoronto.ca as before, the system is setup differently from the previous Teach cluster in the following ways:&lt;br /&gt;
&lt;br /&gt;
* There are now 2 dedicated login nodes, teach-login01 and teach-login02.  &lt;br /&gt;
* The ssh fingerprints for these login nodes can be found on [[Teach_fingerprints]].&lt;br /&gt;
* The compute nodes have 40 cores.  As before, you can request jobs by number of cores.&lt;br /&gt;
* Only temporary lcl_uot.... accounts can log in.&lt;br /&gt;
* Only the home directories of those account are mounted. &lt;br /&gt;
* In particular, the file systems from the other SciNet compute clusters (Niagara, Mist, Trillium,... ) are not and will not be mounted.  You'll need to copy over any files that you need to use on the Teach cluster.&lt;br /&gt;
* There is no $SCRATCH.  You can do all your work on $HOME, which is writable from compute nodes&lt;br /&gt;
* The software stack is the one supplied by the Alliance.  There is no need to load 'CCEnv' to get them.&lt;br /&gt;
* But as before, if you're missing a module, we can still install it for you.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- Do this later&lt;br /&gt;
== Running Jupyter on a Teach Compute Node ==  &lt;br /&gt;
&lt;br /&gt;
1. To be able to run Jupyter on a compute node, you must first (a) install it inside a virtual environment, (b) enable a way for jupyter to seemingly write to a specific directory on $HOME, and (c) create a little helper script called &amp;lt;tt&amp;gt;notebook.sh&amp;lt;/tt&amp;gt; that will be used to start the jupyter server in step 2.  These are the commands that you should use for the installation (which you should do only once, on the Teach login node):&lt;br /&gt;
&lt;br /&gt;
(a) Create virtual env&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
$ module load python/3.9.10&lt;br /&gt;
$ virtualenv --system-site-packages $HOME/.virtualenvs/jupteach&lt;br /&gt;
$ source $HOME/.virtualenvs/jupteach/bin/activate&lt;br /&gt;
$ pip install jupyter jupyterlab&lt;br /&gt;
$ deactivate&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
You can choose another directory than &amp;lt;tt&amp;gt;$HOME/.virtualenvs/jupteach&amp;lt;/tt&amp;gt; for where to create the virtual environment, but you need to be consistent and use the same directory everywhere below.&lt;br /&gt;
&lt;br /&gt;
(b) Make a writable 'runtime' directory for Jupyter. &lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
$ mkdir -p $HOME/.local/share/jupyter/runtime &lt;br /&gt;
$ mv -f $HOME/.local/share/jupyter/runtime $SCRATCH/jupyter_runtime || mkdir $SCRATCH/jupyter_runtime&lt;br /&gt;
$ ln -sT $SCRATCH/jupyter_runtime $HOME/.local/share/jupyter/runtime&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
(c) Create a launch script to use on the compute nodes:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
$ cat &amp;gt; $HOME/.virtualenvs/jupteach/bin/notebook.sh &amp;lt;&amp;lt;EOF&lt;br /&gt;
#!/bin/bash&lt;br /&gt;
source \$HOME/.virtualenvs/jupteach/bin/activate&lt;br /&gt;
export XDG_DATA_HOME=\$SCRATCH/.share&lt;br /&gt;
export XDG_CACHE_HOME=\$SCRATCH/.cache&lt;br /&gt;
export XDG_CONFIG_HOME=\$SCRATCH/.config&lt;br /&gt;
export XDG_RUNTIME_DIR=\$SCRATCH/.runtime&lt;br /&gt;
export JUPYTER_CONFIG_DIR=\$SCRATCH/.config/.jupyter&lt;br /&gt;
jupyter \${1:-notebook} --ip \$(hostname -f) --no-browser --notebook-dir=\$PWD&lt;br /&gt;
EOF&lt;br /&gt;
$ chmod +x  $HOME/.virtualenvs/jupteach/bin/notebook.sh&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2. To run the jupyter server on a compute node, start an interactive session with the &amp;lt;tt&amp;gt;debugjob&amp;lt;/tt&amp;gt; command and then launch the jupyter server:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
$ debugjob -n 16                    # use less if you need less cores.&lt;br /&gt;
$ cd $SCRATCH                       # $HOME is read-only, so move to $SCRATCH&lt;br /&gt;
$ $HOME/.virtualenvs/jupteach/bin/notebook.sh  # add the argument &amp;quot;lab&amp;quot; to start with the jupyter lab&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
Make sure you note down (a) the name of the compute node that you got allocated (they start with &amp;quot;&amp;lt;tt&amp;gt;teach&amp;lt;/tt&amp;gt;&amp;quot; followed by a 2-digit number), and (b) the number following the compute nodes after the colon (usually this is 8888, but it can be another, higher number); this is the PORT  and (c) the last URL that the notebook.sh tells you to use to connect.&lt;br /&gt;
&lt;br /&gt;
4. To connect to this jupyter server running on a teach compute node, which is not accessible from the internet, in a different terminal on you own computer, you must reconnect to the Teach cluster with a port-forwarding&lt;br /&gt;
tunnel to the compute node on which jupyter is running:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
$ ssh -LPORT:teachXX:PORT -o ControlMaster=no USERNAME@teach.scinet.utoronto.ca  -N&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;tt&amp;gt;teachXX&amp;lt;/tt&amp;gt; is to be replaced by the name of the compute node (point (a) above), PORT is to be replaced by the port number that notebook.sh showed, and &amp;lt;tt&amp;gt;USERNAME&amp;lt;/t&amp;gt; should be your teach account username. This command should just &amp;quot;hang&amp;quot; there, it only serves to forward port number PORT (usually 8888) to port PORT (usually 8888) on the compute node.  &lt;br /&gt;
&lt;br /&gt;
Finally, point your browser to the URL that the &amp;lt;tt&amp;gt;notebook.sh&amp;lt;/tt&amp;gt; command printed out (point (b) above), i.e., the one with 127.0.0.1 in it.&lt;br /&gt;
--&amp;gt;&lt;/div&gt;</summary>
		<author><name>Rzon</name></author>
	</entry>
</feed>