<?xml version="1.0" encoding="UTF-8" ?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:media="http://search.yahoo.com/mrss/">
	<channel>
		<title>RCAC - Outages and Maintenance, Announcements, Science Highlights, Events, Coffee Hour Consultations, Outages, Maintenance, Student Events</title>
		<link>https://www.rcac.purdue.edu/news/rss/Radon</link>
		<description><![CDATA[news::news.feed description]]></description>
		<atom:link href="https://www.rcac.purdue.edu/news/rss/Radon" rel="self" type="application/rss+xml" />
		<language>en</language>
		<lastBuildDate>Thu, 16 Apr 2026 00:36:54 EDT</lastBuildDate>
					<item>
				<title><![CDATA[Radon Cluster Retirement]]></title>
				<link>https://www.rcac.purdue.edu/news/836</link>
				<guid isPermaLink="true">https://www.rcac.purdue.edu/news/836</guid>
				<description><![CDATA[<p>The Radon cluster has reached the end of its useful lifespan and will be decommissioned on Monday, April 2nd, 2018.</p>
<p>Radon has served the Purdue research community for a great many years and has undergone radical changes during that time.  However, during that time the landscape of research computing options at Purdue has also changed radically.</p>
<p>The Community Clusters now serve a variety of different research need profiles from traditional HPC to high-memory applications to GPU-enabled code, each of which are expanded or replaced every year. <a href="https://www.rcac.purdue.edu/compute/scholar">Scholar</a> is available for instructional use of research tools within courses. <a href="https://www.rcac.purdue.edu/compute/workbench">Data Workbench</a>, one of our newest offerings, serves those with smaller scale interactive needs.  Radon's mission has been subsumed and is better served by these newer resources.</p>
<p>If you still have need for research computing resources, please contact us at <a href="mailto:rcac-help@purdue.edu">rcac-help@purdue.edu</a> to discuss your options, or you may explore them yourself on this website.  Faculty and staff may directly purchase access to any of the currently available resources from the <a href="https://www.rcac.purdue.edu/purchase">&quot;Purchase&quot; menu</a> on our website as well, by simply selecting what they are looking for and providing a Purdue account number.</p>
<p>We look forward to continuing the development of Research Computing services to meet your evolving needs.  Thank you for your support!</p>
]]></description>
				<pubDate>Mon, 02 Apr 2018 08:00:00 -0400</pubDate>
									<category>Outages and Maintenance</category>
							</item>
					<item>
				<title><![CDATA[All Clusters Outage]]></title>
				<link>https://www.rcac.purdue.edu/news/872</link>
				<guid isPermaLink="true">https://www.rcac.purdue.edu/news/872</guid>
				<description><![CDATA[<p>All Research Computing systems suffered an unplanned outage Saturday, March 24th, 2018 at 8:15pm EDT due to a widespread power failure in the area. Thanks to diligent efforts all night and today by many teams across ITaP, all computational clusters have been returned to normal service as of Sunday, March 25th, 2018 at 4:00pm EDT.</p>
]]></description>
				<pubDate>Sat, 24 Mar 2018 20:15:00 -0400</pubDate>
									<category>Outages</category>
							</item>
					<item>
				<title><![CDATA[New Windows Network Drive (SMB) Access]]></title>
				<link>https://www.rcac.purdue.edu/news/850</link>
				<guid isPermaLink="true">https://www.rcac.purdue.edu/news/850</guid>
				<description><![CDATA[<p>All Windows network drives (SMB/CIFS access) for scratch filesystems on all clusters and home directories have moved!</p>
<p>You should change your mapped network drives to connect to:
<code>\\scratch.my_cluster_name_here.rcac.purdue.edu\my_cluster_name_here</code></p>
<p>Or for home directory access:
<code>\\home.rcac.purdue.edu\home</code></p>
<p>The old server, samba.rcac.purdue.edu, will be retired at the end of this month, so please change your connections as soon as possible to avoid any disruption in service.    This change will allow us to better maintain this type of service during cluster maintenance and upgrades in the future.  More information on this is available in the cluster user guides:</p>
<ul>
<li>
<a href="https://www.rcac.purdue.edu/knowledge/brown/storage/transfer/cifs">Brown</a>
</li>
<li>
<a href="https://www.rcac.purdue.edu/compute/halstead">Halstead</a>
</li>
<li>
<a href="https://www.rcac.purdue.edu/compute/halsteadgpu">HalsteadGPU</a>
</li>
<li>
<a href="https://www.rcac.purdue.edu/compute/rice">Rice</a>
</li>
<li>
<a href="https://www.rcac.purdue.edu/compute/snyder">Snyder</a>
</li>
<li>
<a href="https://www.rcac.purdue.edu/compute/conte">Conte</a>
</li>
<li>
<a href="https://www.rcac.purdue.edu/knowledge/scholar/storage/transfer/cifs">Scholar</a>
</li>
<li>
<a href="https://www.rcac.purdue.edu/compute/radon">Radon</a>
</li>
</ul>
<p>Contact us if you have any questions.  Thank you!</p>
]]></description>
				<pubDate>Thu, 08 Mar 2018 00:00:00 -0500</pubDate>
									<category>Outages and Maintenance</category>
							</item>
					<item>
				<title><![CDATA[Unscheduled Depot Outage on Compute Clusters]]></title>
				<link>https://www.rcac.purdue.edu/news/833</link>
				<guid isPermaLink="true">https://www.rcac.purdue.edu/news/833</guid>
				<description><![CDATA[<p>The servers providing access to Data Depot from  suffered a partial failure.</p>
<p>Many nodes in these clusters temporarily lost access to Depot. Jobs accessing files on Depot may have paused or terminated depending on the severity of the issue on each node and the nature of the job being run.</p>
<p>The issue is already being corrected and access restored, but some job failures or delays may be seen.</p>
]]></description>
				<pubDate>Wed, 03 Jan 2018 11:30:00 -0500</pubDate>
									<category>Outages</category>
							</item>
					<item>
				<title><![CDATA[Holiday Break]]></title>
				<link>https://www.rcac.purdue.edu/news/817</link>
				<guid isPermaLink="true">https://www.rcac.purdue.edu/news/817</guid>
				<description><![CDATA[<p>Purdue University will be observing a holiday break from December 23 - January 2.</p>
<p>During this time, Research Computing services will continue to be available, but all staff will be on leave.  Critical system outages will be dealt with should they occur during this time.  Other support needs will be handled when the University re-opens on January 3rd.  There will also be no coffee hour consultations during this break.</p>
<p><strong>Scratch file purging (only on community clusters with scratch space) will continue as normal during the break, so be sure to archive your files in scratch storage or check for any upcoming purge warning email during the break. This does not apply to Data Depot or home directories--only scratch storage.</strong></p>
<p>Have a wonderful break, everyone, and we look forward to great things in the new year!</p>
]]></description>
				<pubDate>Sat, 23 Dec 2017 00:00:00 -0500</pubDate>
									<category>Announcements</category>
							</item>
					<item>
				<title><![CDATA[Radon, Hathi, and WSC Hadoop Cluster Maintenance]]></title>
				<link>https://www.rcac.purdue.edu/news/791</link>
				<guid isPermaLink="true">https://www.rcac.purdue.edu/news/791</guid>
				<description><![CDATA[<p>The Hathi, Radon, and WSC Hadoop clusters will be unavailable beginning at Friday, November 3rd, 2017 at 5:00pm EDT, for scheduled electrical power maintenance. The clusters will return to full production by Friday, November 3rd, 2017 at 11:59pm EDT.</p>
<p>During this time, some of the power feeding Hathi, Radon, and WSC Hadoop will be upgraded.</p>
<p>Any PBS jobs which request a walltime which would take them past Friday, November 3rd, 2017 at 5:00pm EDT will not start and will remain in the queue until after the maintenance is completed.</p>
]]></description>
				<pubDate>Fri, 03 Nov 2017 17:00:00 -0400</pubDate>
									<category>Maintenance</category>
							</item>
					<item>
				<title><![CDATA[Unscheduled Outage in Math Data Center]]></title>
				<link>https://www.rcac.purdue.edu/news/767</link>
				<guid isPermaLink="true">https://www.rcac.purdue.edu/news/767</guid>
				<description><![CDATA[<p>At approximately 2:00pm EDT on Tuesday, September 5th, 2017, the Math building data center lost some power feeds which supply the Conte, Halstead, HalsteadGPU, Hathi, and Radon clusters. Scheduling on these has been paused until we can be sure power is stable again. It appears only some portions of these systems were affected.</p>
<p>We will update this posting as we know more.</p>
]]></description>
				<pubDate>Tue, 05 Sep 2017 14:00:00 -0400</pubDate>
									<category>Outages</category>
							</item>
					<item>
				<title><![CDATA[Scratch Purge Policy Change]]></title>
				<link>https://www.rcac.purdue.edu/news/734</link>
				<guid isPermaLink="true">https://www.rcac.purdue.edu/news/734</guid>
				<description><![CDATA[<p>After a long review, Research Computing has determined it is necessary to alter the scratch storage purge policy on all systems.</p>
<p><em><strong>Effective August 28, 2017, all scratch storage systems will begin purging files which have not been accessed (for either reading or writing) in 60 days or more on a weekly basis.</strong></em></p>
<p>This is a change from the previous policy of 90 days.  This change is necessary in order to maintain a minimal level of free space on our scratch storage systems and ensure they remain effective for everyone.</p>
<p>Any files which are in active use by jobs on the clusters will not be purged, nor will you need to take any steps to ensure they are not.  Please refrain from explicitly taking actions to exempt files from purging on a long-term basis, as this impedes our ability to offer the generous spaces we do for all users. It may also endanger your research, because there are no backups or recovery options for lost files in scratch storage spaces.</p>
<p>As soon as you are no longer actively computing against data, we strongly recommend copying that data to either a Data Depot space or the Fortress Archive, where your data will be safe for any future needs. There is information on how to access both of these systems on our website, and you may contact us at <a href="mailto:rcac-help@purdue.edu">rcac-help@purdue.edu</a> at any time with questions. If you need personal assistance, our staff also hold <a href="https://www.rcac.purdue.edu/news/545">Coffee Break Consultation </a>.</p>
<p>Thank you!</p>
]]></description>
				<pubDate>Thu, 03 Aug 2017 00:00:00 -0400</pubDate>
									<category>Announcements</category>
							</item>
					<item>
				<title><![CDATA[Hathi, Radon, and Specialized Cluster Maintenance]]></title>
				<link>https://www.rcac.purdue.edu/news/725</link>
				<guid isPermaLink="true">https://www.rcac.purdue.edu/news/725</guid>
				<description><![CDATA[<p><strong>Update message:</strong>
After performing necessary repairs, Radon has been returned to service.</p>
<p>--
<strong>Previous message:</strong>
After consulting with vendor support, we have determined that Radon has experienced a failure in its network hardware.  Parts and and a vendor engineer will be on-site Tuesday, June 27  to repair. Radon will remain unavailable until that time.</p>
<p>--
<strong>Previous message:</strong></p>
<p>The outage is being extended until 11:59am.  There have been continued issues bringing Radon back into production due to an issue with its connectivity.  At the moment, it cannot be said whether this is due to a networking problem or an issue within the Radon internal server configuration.</p>
<p>As previously stated, all other resources are back online and in production.</p>
<p>Update by 11:59a.</p>
<p>Thank you for your continued patience.</p>
<hr />
<p>We are extending the outage by two hours due to issues encountered with bringing the Radon cluster back into production.</p>
<p>All other resources are back in online and serving at this time.</p>
<p>Will update by 2:00am with status.</p>
<p>ORIGINAL OUTAGE NOTICE:</p>
<p>The  clusters, as well as some restricted access highly-specialized resources will be unavailable beginning at Friday, June 23rd, 2017 at 5:00pm EDT, for scheduled power maintenance. The clusters will return to full production by Tuesday, June 27th, 2017 at 2:00pm EDT.</p>
<p>During this time, power work will be performed in the Math data center.</p>
<p>Any PBS jobs which request a walltime on Radon which would take them past Friday, June 23rd, 2017 at 5:00pm EDT will not start and will remain in the queue until after the maintenance is completed.  The other systems affected do not use schedulers which are directly aware of this upcoming maintenance.</p>
]]></description>
				<pubDate>Fri, 23 Jun 2017 17:00:00 -0400</pubDate>
									<category>Maintenance</category>
							</item>
					<item>
				<title><![CDATA[Removal of netcdf and hdf5 system-level library installations]]></title>
				<link>https://www.rcac.purdue.edu/news/740</link>
				<guid isPermaLink="true">https://www.rcac.purdue.edu/news/740</guid>
				<description><![CDATA[<p>Research Computing has begun removing several libraries installed at the system-level that should be provided by the module command. These libraries include netcdf, hdf5, and several related packages. This change should have limited impact as module installations of these packages are not changing and most people should be using the module versions.</p>
<p>These packages are being removed at the system-level because these types of libraries should be managed at the &quot;module&quot; level. These packages were inadvertently installed but have not been tuned and optimized for use on the clusters. Some software compilations may have accidentally picked up these packages, rather than the &quot;module&quot; package, or may have picked up some combination of system-level packages and module packages. Anyone with software linked against these libraries may have had issues with the software mixing versions.</p>
<p>Many scientific applications are very sensitive to the compilers used to build software and its libraries, so by removing these packages at the system-level, and by the &quot;module&quot; command's enforcement of only loading one of a package at the time, you can be sure your code is being compiled with exactly the same version of the library and not scrambled with other versions that were not asked for.</p>
<p>There were a couple of our own module installations that fell victim to this scrambling of libraries which we have fixed. If your software does not use netcdf or hdf5, or your software does and you are loading a netcdf or hdf5 module, you will not be impacted. If you see an error such as:</p>
<pre><code>error while loading shared libraries: [libnetcdf.so](http://libnetcdf.so).6: cannot open shared object file: No such file or directory
</code></pre>
<p>You will need to ensure you are loading a netcdf or hdf5 module and you may need to recompile your code to ensure you are picking up the library version from the module.</p>
<p>Please let us know at <a href="mailto:rcac-help@purdue.edu">rcac-help@purdue.edu</a> if you have any questions.</p>
]]></description>
				<pubDate>Wed, 14 Jun 2017 00:00:00 -0400</pubDate>
									<category>Announcements</category>
							</item>
					<item>
				<title><![CDATA[Data Depot Outage]]></title>
				<link>https://www.rcac.purdue.edu/news/720</link>
				<guid isPermaLink="true">https://www.rcac.purdue.edu/news/720</guid>
				<description><![CDATA[<p>Engineers have restored failed core servers back to a functional state. Data Depot is up and running as normal and job scheduling resumed. Should you encounter any lingering issues please let us know at <a href="mailto:rcac-help@purdue.edu">rcac-help@purdue.edu</a></p>
<p><strong>Original Message</strong></p>
<p>Some core servers which operate the Research Data Depot have suffered a failure.  Engineers are working to correct this now.  To prevent any new jobs from starting while Depot is not functioning normally, scheduling on all major clusters has been paused.</p>
<p>We will update this as we learn more about the nature of the issue.</p>
]]></description>
				<pubDate>Thu, 18 May 2017 13:50:00 -0400</pubDate>
									<category>Outages</category>
							</item>
					<item>
				<title><![CDATA[Clusters to complete transition to hierarchy modules]]></title>
				<link>https://www.rcac.purdue.edu/news/700</link>
				<guid isPermaLink="true">https://www.rcac.purdue.edu/news/700</guid>
				<description><![CDATA[<p>The transition to hierarchy modules has completed today, May 9th. If any of your job scripts still have old module names <strong>the module load will no longer work</strong> so be sure to double check your PBS job scripts and output for any old module names. We have attempted to personally address individuals using the old names based on usage logs over the past month, and will do so again today. For more information about hierarchy module see <a href="https://www.rcac.purdue.edu/news/924">the original news article</a>. If you have any questions about this please contact us at <a href="mailto:rcac-help@purdue.edu">rcac-help@purdue.edu</a>.</p>
<p><strong>Original Message</strong></p>
<p>On May 9, 2017  - December 31, 2017  the clusters will complete the change to hierarchy modules. When the clusters switched to hierarchy modules on February 1st, the old module names remained supported with a warning message in place. <strong>On May 9, 2017  - December 31, 2017 , support for this translation of old module names will end.</strong></p>
<p>You should check PBS job output for warning messages about old module style names. Any 'module load' commands showing this warning will fail after May 9, 2017  - December 31, 2017 . The warning message will tell you what you should change your 'module load' commands to.</p>
<p>You can switch over early with the command:</p>
<pre><code>touch ~/.hierarchy
</code></pre>
<p>This will disable the automatic translation and will put you in and environment equivalent to that of after May 9, 2017  - December 31, 2017 .</p>
<p>For more information about hierarchy module see <a href="https://www.rcac.purdue.edu/news/924">the original news article</a>. If you have any questions about this please contact us at <a href="mailto:rcac-help@purdue.edu">rcac-help@purdue.edu</a>.</p>
]]></description>
				<pubDate>Tue, 09 May 2017 00:00:00 -0400</pubDate>
									<category>Announcements</category>
							</item>
					<item>
				<title><![CDATA[Emergency Security Patching of RCAC Clusters]]></title>
				<link>https://www.rcac.purdue.edu/news/668</link>
				<guid isPermaLink="true">https://www.rcac.purdue.edu/news/668</guid>
				<description><![CDATA[<p>Due to a recent security vulnerability, the  clusters will have their operating system upgraded to a newer version  during February 2, 2017  5:00pm - March 2, 2017  5:00pm EST. Unlike other cluster downtimes, this upgrading process will follow a &quot;rolling reboot&quot; strategy, that is, nodes will be updated and rebooted when jobs currently running on them complete.</p>
<p><strong>Potential impact on users:</strong></p>
<ul>
<li>Currently running batch jobs will NOT be impacted.</li>
<li>For each front-end server, users will be given a 48-hour notice to save their work and exit from any currently running interactive jobs. Interactive jobs still running at the time of reboot will be terminated.</li>
<li>Users may experience slightly longer scheduling delays during initial hours of the updating process.</li>
</ul>
]]></description>
				<pubDate>Thu, 02 Feb 2017 17:00:00 -0500</pubDate>
									<category>Maintenance</category>
							</item>
					<item>
				<title><![CDATA[Radon to switch to hierarchical modules]]></title>
				<link>https://www.rcac.purdue.edu/news/652</link>
				<guid isPermaLink="true">https://www.rcac.purdue.edu/news/652</guid>
				<description><![CDATA[<p>Radon has been converted to hierarchical modules.</p>
<p>Old style module load commands should still continue to function until the end of the Spring semester, however you will be given a warning message and a suggestion on how to correct your scripts. Once you believe you have made all the necessary changes you may turn off the old-name translation by:</p>
<pre><code>$ touch ~/.hierarchy
</code></pre>
<p>Then, by logging out and back in, the automatic translation will be disabled. This will allow you to be sure that your scripts are correctly updated before the translation is turned off system wide at the end of the semester. We will send further details and timing on turning off translations in the coming weeks.</p>
<p><strong>Original Message</strong></p>
<p>On January 25th, 2017 the Radon cluster's software stack will be converted to an hierarchical configuration. This new software has been in use on <a href="https://www.rcac.purdue.edu/news/911">Halstead since it was brought online</a>. This change will help prevent errors and configuration problems due to version mismatches or conflicting software, and will allow for ITaP to provide a more robust software stack.</p>
<p>This change may require you to make small modifications to your job scripts. The module command will attempt to automatically translate these for you to help ease into the transition but an automatic translation may not always be possible, and this functionality will only be offered for a limited time.</p>
<p>To further aid in this transition, a Radon node has been converted to the new configuration for you to log in and explore the new configuration. In addition, a small tool has been developed that will scan your shell startup scripts and specified job scripts to identify potential changes that may need to be made.</p>
<p>From a Radon front-end machine, type the command:</p>
<pre><code>$ ssh radon-modules
</code></pre>
<p>Once on the test node, you can use &quot;module spider&quot; and &quot;module avail&quot; to explore the new hierarchical configuration. Once on this test node you can run the module checker script.</p>
<p>To check shell start up scripts:</p>
<pre><code>$ modulecheck
</code></pre>
<p>To check a job script(s):</p>
<pre><code>$ modulecheck myjobscript.sub
</code></pre>
<p>The tool will attempt to also scan any other scripts that the specified scripts source or include. If the tool is unable to find a sourced file, it will warn you that it was unable to check that file.</p>
<p>The output of this command will give you a summary of the files it found, whether it found any issues and will offer suggestions on how to update your scripts. Please do not try to run your jobs on this test node.</p>
<p>Additional information and explanations about the module hierarchy can be found in <a href="https://www.rcac.purdue.edu/compute/radon">the user guide</a>.</p>
<p>If you have any questions or concerns please contact us at <a href="mailto:rcac-help@purdue.edu">rcac-help@purdue.edu</a>. The rest of the ITaP Research Computing clusters will follow in conversion <a href="https://www.rcac.purdue.edu/news/924">a week later on February 1st, 2017</a>.</p>
]]></description>
				<pubDate>Wed, 25 Jan 2017 00:00:00 -0500</pubDate>
									<category>Announcements</category>
							</item>
					<item>
				<title><![CDATA[Job scheduling paused on Radon]]></title>
				<link>https://www.rcac.purdue.edu/news/638</link>
				<guid isPermaLink="true">https://www.rcac.purdue.edu/news/638</guid>
				<description><![CDATA[<p>Job scheduling was paused on Radon between 6 pm and 7 pm this evening. Node monitoring processes marked most nodes offline around 6 pm, preventing new jobs from starting. System engineers cleared the fault in the node monitoring, and nodes came back online at around 7pm.</p>
<p>No impact to existing jobs is suspected, but new jobs would not have started during this time. If you see any issues, please let us know at <a href="mailto:rcac-help@purdue.edu">rcac-help@purdue.edu</a></p>
]]></description>
				<pubDate>Mon, 14 Nov 2016 18:00:00 -0500</pubDate>
									<category>Outages</category>
							</item>
					<item>
				<title><![CDATA[Security vulnerability patch impacts debugging]]></title>
				<link>https://www.rcac.purdue.edu/news/622</link>
				<guid isPermaLink="true">https://www.rcac.purdue.edu/news/622</guid>
				<description><![CDATA[<p>Due to a recently found vulnerability in the Linux Kernel (known as the Dirty-COW vulnerability), an emergency patch has been applied on the cluster nodes. This patch is necessary to avoid exploitation of the vulnerability. Unfortunately, the patch impacts/disables the Linux kernel's &quot;ptrace&quot; functionality. Tools that use &quot;ptrace&quot; (including Totalview and Intel vTune) are affected by this.</p>
<p>A complete fix of the vulnerability will require an upgrade of the Operating System on the cluster nodes. We are currently working on scheduling a downtime for the clusters to perform this upgrade.</p>
]]></description>
				<pubDate>Mon, 31 Oct 2016 00:00:00 -0400</pubDate>
									<category>Announcements</category>
							</item>
					<item>
				<title><![CDATA[Home Filesystem Maintenance - All Clusters]]></title>
				<link>https://www.rcac.purdue.edu/news/568</link>
				<guid isPermaLink="true">https://www.rcac.purdue.edu/news/568</guid>
				<description><![CDATA[<p>Conte has been returned to normal operations as well now.  This concludes the home directory maintenance on all systems.</p>
<p><strong>Update: September 27, 2016  11:55pm</strong></p>
<p>All systems other than Conte have been successfully returned to normal operations with the new home directory filesystem.  Work continues at this point on Conte to ensure the Phi accelerators are properly reconfigured.</p>
<p>Carter has also been given a new scratch filesystem during this maintenance.  This should alleviate some of the problems with the previous scratch filesystem on Carter.  For more details, please see the Carter-specific announcement on this topic:  <a href="https://www.rcac.purdue.edu/news/887">New Carter Scratch Filesystem</a></p>
<p><strong>Reminder:</strong></p>
<p>This is a reminder of the Home Filesystem Maintenance taking place next week on Tuesday, September 27th.</p>
<p>Details below.</p>
<p><strong>Original Message:</strong></p>
<p>All of the research clusters (Carter, Coates, Hansen, Radon, and Rossmann) as well as some other minor systems will be unavailable beginning at Tuesday, September 27th, 2016 at 7:00am EDT, for scheduled maintenance. All clusters other than Conte will return to full production by 11:59pm.</p>
<p>Conte will return to partial capacity by that time, but will not return to full production until the following day. Many Conte nodes will remain offline and gradually be returned to service over the following 12-24 hours to allow for power reconfiguration in the data center.  Please see the separate article on Conte: <a href="https://www.rcac.purdue.edu/news/873">Conte Cluster Maintenance</a>.</p>
<p>During the large all-systems maintenance Tuesday, the /home filesystem used by all Research Computing systems will be replaced by a new filesystem.  The new filesystem will be based on DDN's GRIDScalar technology and running on new hardware dedicated exclusively to Research Computing home directories.</p>
<p>All files on the existing /home filesystem will be migrated to the new system during the maintenance window and prior to any of the clusters returning to service.</p>
<p>In the coming weeks, any jobs which request a walltime which would take them past Tuesday, September 27th, 2016 at 7:00am EDT will not start and will remain in the queue until after the maintenance is completed.</p>
]]></description>
				<pubDate>Tue, 27 Sep 2016 07:00:00 -0400</pubDate>
									<category>Maintenance</category>
							</item>
					<item>
				<title><![CDATA[Software stack changes and upgrades]]></title>
				<link>https://www.rcac.purdue.edu/news/593</link>
				<guid isPermaLink="true">https://www.rcac.purdue.edu/news/593</guid>
				<description><![CDATA[<p>During the <a href="https://www.rcac.purdue.edu/news/871">Home Filesystem Maintenance - All Clusters</a> maintenance on September 27th, several upgrades and changes will be made to the software stack on the clusters. Changes will include updates to the default version of the Intel compiler and associated software stack as well to the default MPI libraries. Some older versions of other software will also be removed. These changes are being made in order to bring clusters in line with the software environment that is being planned for the new Halstead cluster.</p>
<p>These upgrades will provide the best performance for the new and existing clusters and will provide a consistent Intel version stack across all of our clusters. The new software stack is currently available on the clusters for testing and upgrade. ITaP research computing staff recommends testing out the new compilers and upgrading prior to September 27th.</p>
<p>WHAT WILL BE THE IMPACT TO INTEL COMPILERS?</p>
<p>We will be upgrading the default Intel version from 13.1.1.163 to 16.0.1.150. The current default has been around for several years, and many researchers are already switching to the latest versions of Intel compilers. The 13.1.1.163 version will remain available on the current clusters for a period of time to give researchers time to finish up projects and upgrade to the latest. Any software dependent on the default version of Intel 13.1.1.163 will also have it's default upgraded.</p>
<p>WHAT WILL BE THE IMPACT TO MPI LIBRARIES?</p>
<p>We will be upgrading the default version of OpenMPI from 1.6.3 to 1.8.1. This new versions offers stability and performance enhancements and some new features. Version 1.8.1 has been available for some time and many researchers have already moved to 1.8.1.</p>
<p>We will be upgrading the default version of IMPI from 4.1.1.036 to 5.1.2.150. This new versions offers stability and performance enhancements and some new features. Version 5.1.2.150 has been available for some time and many researchers have already moved to 5.1.2.150.</p>
<p>Any software dependent on one of these default MPI versions will also have it's default upgraded appropriately.</p>
<p>It is recommended that you upgrade to these new libraries, however, if you need to continue using the old default versions you may do so by switching your &quot;module load&quot; to the specific version. The Intel 13 stack will remain available for those who require it. These new compilers offer bug fixes and enhanced performance and stability. Users are encouraged to send in any experiences with these new compilers to help us evaluate the direction of new compilers on RCAC systems.</p>
<p>WHAT OTHER SOFTWARE WILL BE IMPACTED?</p>
<p>There will be several changes to other miscellaneous software. Older versions of some software will be removed in favor of newer versions. Default versions of a few software will be updated to the latest version. In most cases, these older versions are being infrequently used so most should not be impacted by these changes.</p>
<p>If any software you are using will be impacted by these changes you will see a notice message being printed to your session or in your job output files when loading an affected module. This notice will provide recommendations on the latest version.</p>
<p>HOW DO I KNOW IF MY WORKFLOW WILL BE IMPACTED?</p>
<p>Whenever a module that will be impacted is loaded a notice is printed to your screen or job output log. Please take a look at your job output over the next couple of weeks and make note of any changes being advertised. You may continue using these modules as-is until September 27th to allow time to make any changes necessary. Users are encouraged to make any changes necessary beforehand to avoid disruption when changes are made.</p>
<p>WHAT IF AN IMPACTED MODULE IS REQUIRED BY MY RESEARCH?</p>
<p>We understand some users may not be able to change compilers or MPI libraries in the middle of a research project. Modules involved in a default version update will continue to be available, however, you will need to update your job scripts to request the specific version of the module. If you are already loading specific versions no changes are necessary.</p>
<p>If a version of software you depend on is being completely removed and you are unable to upgrade, please contact us at <a href="mailto:rcac-help@purdue.edu">rcac-help@purdue.edu</a>. We will help you transition to a newer version if possible, or provide you with a copy of the old software version.</p>
<p>WHY ARE YOU CHANGING THE SOFTWARE STACK?</p>
<p>ITaP aims to provide a software stack that allows for optimal use (performance and stability) of the clusters. This necessitates periodic updates to the stack as compilers, libraries, and software are improved over time. By removing older modules from the main stack we help ensure the selection is simple and easy for users to find the best compilers and libraries to use. If no modules were removed the selection would become difficult to navigate as well become difficult for ITaP staff to manage. Any major changes will be coordinated with scheduled maintenance periods to minimize impact.</p>
<p>If you have any questions or concerns with the upcoming changes please contact <a href="mailto:rcac-help@purdue.edu">rcac-help@purdue.edu</a></p>
]]></description>
				<pubDate>Tue, 27 Sep 2016 00:00:00 -0400</pubDate>
									<category>Announcements</category>
							</item>
					<item>
				<title><![CDATA[Degraded performance of several systems]]></title>
				<link>https://www.rcac.purdue.edu/news/605</link>
				<guid isPermaLink="true">https://www.rcac.purdue.edu/news/605</guid>
				<description><![CDATA[<p>We have seen a significant wave of these events this morning, September 21.  For the most part, this wave seems to have been linked to a storage problem that has been resolved.  However, we are implementing new monitoring and response procedures today to ensure a similar recurrence is caught and dealt with much more quickly.</p>
<p><strong>Original Message:</strong></p>
<p>System, Network, Storage, and Support staff are working to diagnose and correct issues that have been seen recently within ITaP's Research Computing systems.</p>
<p>Symptoms being reported involve an apparent complete freeze of open sessions, the inability to open new login sessions, difficulties using text editors, and disruptions in file access. In cases we have seen, these events seem to last for about 3-5 minutes, then clear up.  However, there may be ongoing effects on jobs running on the Research Clusters, including job failure due to the storage access disruption.</p>
<p>We are examining log files and monitoring processes actively, and are working to correlate the timing of these events across our systems, and expect to identify a fundamental cause that we can then correct.  At this time, however, we do not have an estimated time for a fix.</p>
<p>Please follow this news item for further information.</p>
]]></description>
				<pubDate>Tue, 13 Sep 2016 00:00:00 -0400</pubDate>
									<category>Outages</category>
							</item>
					<item>
				<title><![CDATA[Radon Cluster Maintenance]]></title>
				<link>https://www.rcac.purdue.edu/news/557</link>
				<guid isPermaLink="true">https://www.rcac.purdue.edu/news/557</guid>
				<description><![CDATA[<p><strong>Update</strong>
As of 5:50 pm, Tuesday, 16 Aug 2016, the Radon cluster has been returned to service and is fully operational.</p>
<p>Thank you for your patience.</p>
<p><strong>Update</strong></p>
<p>Due to unanticipated conflicts between the upgraded scheduler and our network configuration, the outage for Radon has been extended,  while we work to eliminate the conflict.</p>
<p>We will have a further update no later than 7:00 pm.</p>
<p><strong>Original Message</strong></p>
<p>The  cluster will be unavailable beginning at Tuesday, August 16th, 2016 at 8:00am EDT, for scheduled maintenance. The cluster will return to full production by Tuesday, August 16th, 2016 at 6:00pm EDT.</p>
<p>During this time,  will have the operating system patched, the PBS resource management system upgraded, and maintenance performed on the Lustre scratch filesystem.</p>
<p>Any PBS jobs which request a walltime which would take them past Tuesday, August 16th, 2016 at 8:00am EDT will not start and will remain in the queue until after the maintenance is completed.</p>
]]></description>
				<pubDate>Tue, 16 Aug 2016 08:00:00 -0400</pubDate>
									<category>Maintenance</category>
							</item>
			</channel>
</rss>