Difference between revisions of "Main Page"

From SciNet Users Documentation
Jump to: navigation, search
(System Status)
(System Status)
 
(276 intermediate revisions by 10 users not shown)
Line 9: Line 9:
 
|{{Up|Niagara|Niagara_Quickstart}}
 
|{{Up|Niagara|Niagara_Quickstart}}
 
|{{Up|HPSS|HPSS}}
 
|{{Up|HPSS|HPSS}}
|{{Up|SOSCIP GPU|SOSCIP_GPU}}
+
|{{Up|Mist|Mist}}
|{{Up|P8|P8}}
+
|{{Up|Teach|Teach}}
 
|-
 
|-
|{{Up|Teach|Teach}}
 
 
|{{Up|Jupyter Hub|Jupyter_Hub}}
 
|{{Up|Jupyter Hub|Jupyter_Hub}}
 
|{{Up|Scheduler|Niagara_Quickstart#Submitting_jobs}}
 
|{{Up|Scheduler|Niagara_Quickstart#Submitting_jobs}}
 
|{{Up|File system|Niagara_Quickstart#Storage_and_quotas}}
 
|{{Up|File system|Niagara_Quickstart#Storage_and_quotas}}
 +
|{{Up|Burst Buffer|Burst_Buffer}}
 
|-
 
|-
 +
|{{Up|Login Nodes|Niagara_Quickstart#Logging_in}}
 
|{{Up|External Network|Niagara_Quickstart#Logging_in}}  
 
|{{Up|External Network|Niagara_Quickstart#Logging_in}}  
 
|{{Up|Globus|Globus}}
 
|{{Up|Globus|Globus}}
 
|}
 
|}
 +
 
<!-- Current Messages: -->
 
<!-- Current Messages: -->
  
<b> Sun 30 Jun 30 2019 </b> The <b>SOSCIP BGQ</b> and <b>P7</b> systems were decommissioned on <b>June 30th, 2019</b>. The BGQdev front end node and storage are still available.   
+
From Tue Mar 30 at 12 noon EST to Thu Apr 1 at 12 noon EST, there will be a two-day reservation for the "Niagara at Scale" pilot eventDuring these 48 hours, only "Niagara at Scale" projects will run on the compute notes (as well as SOSCIP projects, on a subset of nodes). All other users can still login, access their data, and submit jobs throughout this event, but the jobs will not run until after the eventThe debugjob queue will remain available to  everyone as well.
 
 
<b>Wed 19 Jun 2018, 1:20:00 PM:</b> The BGQ is back online.
 
 
 
<b>Wed 19 Jun 2018, 10:00:00 AM:</b> The BGQ is still down, the SOSCIP GPU nodes should be back up.  
 
 
 
<b>Wed 19 Jun 2018, 1:40:00 AM:</b> There was an issue with the SOSCIP BGQ and GPU Cluster last night about 1:42am, probably a power fluctuation that took it down.   
 
 
 
<b>Wed 12 Jun 2019, 3:30 AM - 7:40 AM</b> Intermittent system issues on Niagara's project and scratch as the file number limit was reached. We increased the number of files allowed in total on the file system.  
 
  
<b>Thu 30 May 2019, 11:00:00 PM:</b>
+
The scheduler will not start batch jobs that cannot finish before the start of this event. Users can submit small and short jobs can take advantage of this, as the scheduler may be able to fit these jobs in before the event starts on the otherwise idle nodes.
Niagara maintenance finished. Some action on the part of users will be required when they first connect again to a Niagara login nodes or datamovers. This is due to the security upgrade of the Niagara cluster, which is now in line with currently accepted best practices.  The details of the required actions can be found on the [[SSH Changes in May 2019]] wiki page.
 
  
<b>Fri 5 Apr 2019</b> Software updates on Niagara: The default CCEnv software stack now uses avx512 on Niagara, and there is now a NiaEnv/2019b stack ("epoch").  
+
Tue 23 Mar 2021 12:19:07 PM EDT - Planned external network maintenance 12pm-1pm Tuesday, March 23rd.  
  
<b>Thu 4 Apr 2019</b> The 2019 compute and storage allocations have taken effect on Niagara.
 
 
<!--  When removing system status entries, please archive them to: https://docs.scinet.utoronto.ca/index.php/Previous_messages -->
 
<!--  When removing system status entries, please archive them to: https://docs.scinet.utoronto.ca/index.php/Previous_messages -->
 
{|style="border-spacing: 10px;width: 100%"
 
{|style="border-spacing: 10px;width: 100%"
Line 45: Line 37:
 
* [[Niagara Quickstart]]
 
* [[Niagara Quickstart]]
 
* [[HPSS | HPSS archival storage]]
 
* [[HPSS | HPSS archival storage]]
* [[BGQ | SOSCIP BlueGene/Q cluster]]
+
* [[Mist| Mist Power 9 GPU cluster]]
* [[SOSCIP_GPU | SOSCIP GPU cluster]]
 
* [[P7|Experimental Power 7 cluster]]
 
* [[P8|Experimental Power 8 GPU cluster]]
 
 
* [[Teach|Teach cluster]]
 
* [[Teach|Teach cluster]]
 
* [[FAQ | FAQ (frequently asked questions)]]
 
* [[FAQ | FAQ (frequently asked questions)]]
Line 55: Line 44:
  
 
== Tutorials, Manuals, etc. ==
 
== Tutorials, Manuals, etc. ==
* [https://courses.scinet.utoronto.ca SciNet education material]
+
* [https://support.scinet.utoronto.ca/education/browse.php SciNet education material]
 
* [https://www.youtube.com/c/SciNetHPCattheUniversityofToronto SciNet's YouTube channel]
 
* [https://www.youtube.com/c/SciNetHPCattheUniversityofToronto SciNet's YouTube channel]
 
* [[Modules specific to Niagara|Software Modules specific to Niagara]]  
 
* [[Modules specific to Niagara|Software Modules specific to Niagara]]  
Line 61: Line 50:
 
* [[Burst Buffer]]
 
* [[Burst Buffer]]
 
* [[SSH Tunneling]]
 
* [[SSH Tunneling]]
 +
* [[SSH#Two-Factor_authentication|Two-Factor Authentication]]
 
* [[Visualization]]
 
* [[Visualization]]
 
* [[Running Serial Jobs on Niagara]]
 
* [[Running Serial Jobs on Niagara]]
 
* [[Jupyter Hub]]
 
* [[Jupyter Hub]]
 
|}
 
|}

Latest revision as of 17:32, 1 April 2021

System Status

Niagara HPSS Mist Teach
Jupyter Hub Scheduler File system Burst Buffer
Login Nodes External Network Globus


From Tue Mar 30 at 12 noon EST to Thu Apr 1 at 12 noon EST, there will be a two-day reservation for the "Niagara at Scale" pilot event. During these 48 hours, only "Niagara at Scale" projects will run on the compute notes (as well as SOSCIP projects, on a subset of nodes). All other users can still login, access their data, and submit jobs throughout this event, but the jobs will not run until after the event. The debugjob queue will remain available to everyone as well.

The scheduler will not start batch jobs that cannot finish before the start of this event. Users can submit small and short jobs can take advantage of this, as the scheduler may be able to fit these jobs in before the event starts on the otherwise idle nodes.

Tue 23 Mar 2021 12:19:07 PM EDT - Planned external network maintenance 12pm-1pm Tuesday, March 23rd.

QuickStart Guides

Tutorials, Manuals, etc.