#

Blog

Boston Data Center Move – Jan. 21-23, 2025 (Phase 2)

FAS Research Computing will be moving floors in our Boston data center in early 2025. This will require at least two downtimes to accomplish. The first phase will be completed Jan 12-13, 2025. Phase 2 will take place Jan 21-23 and will be less disruptive and will not require a cluster shutdown. Phase 2: The cluster will be online. Some…

Boston Data Center Move – Jan. 12-14, 2025 (Phase 1)

FAS Research Computing will be moving floors in our Boston data center in early 2025. This will require at least two downtimes to accomplish. The most impactful will be the first move of our Boston Isilon storage which houses all user home directories.  Phase 1: The cluster will be offline until Tuesday 1/14. This first move will begin at 4PM …

Announcing Netscratch: Our New High-Performance Scratch Filesystem 

This was sent as an email(s) to all cluster users on Nov 1st, 2024. It has been edited here to provide additional information about moving data. We’re excited to introduce netscratch, a high-performance scratch filesystem powered by Vast Data. After thorough evaluation, Vast Data’s advanced flash technology stood out for its performance and scalability, making it our solution of choice.   Data…

2024 MGHPCC power downtime May 21-24, 2024

POWERUP COMPLETE 6:20PM See Status Page   The yearly power downtime at our Holyoke data center, MGHPCC, has been scheduled.  This year the power outage will take place on a Wednesday/Thursday the week before the Memorial Day long weekend (Mon. 5/27/24). Ordinarily this is happens during a Monday/Tuesday or Tuesday/Wednesday timeframe, but due to scheduling conflicts with other universities that…

Cannon 2.0

Summary FASRC is adding 216 Intel Sapphire Rapids nodes with 1TB of RAM each, 4 Intel Sapphire Rapids nodes with 2TB of RAM each and 144 A100 80GB GPUs to the Cannon cluster. The Sapphire Rapids cores will be made available in the new ‘sapphire’ partition. The new A100 GPUs will be added to the ‘gpu’ partition.  Partitions will be…

2023 MGHPCC downtime, and major OS and software changes

DOWNTIME COMPLETE The annual multi-day power downtime at MGHPCC (https://www.rc.fas.harvard.edu/blog/2023-downtime/) is complete (with exceptions noted below). Normal service resumes today (Friday June 9th) at 9am. As noted in previous communications, many changes have been made to the cluster and software. If you run jobs on the cluster and did not previously try out the test cluster, you will need to make…

Cluster Fragmentation

by Paul Edmon June 3, 2022 A common pitfall of High Performance Computing (HPC) scheduling is cluster fragmentation.  This is not unique to HPC mind you, any system where you have a limited amount of space that you try to fill with quasi random sized blocks of stuff will end up fragmented at some level (see your last game of…

2021 Cluster Upgrades – Sept. 30th 2021

  We have some exciting news about new resources we want to share with you. In the summer of 2019 we invested in direct water cooling and were able to bring you the Cannon cluster with 32,000 of the fastest Intel Cascade Lake cores. This 8-rack block of compute has formed the core of our computational capacity for the past…

Security advisory regarding Python/Conda/pip/PyPI

AUDIENCE: All Python/Conda users IMPACT: Potential malicious packages installed or malware downloaded Numerous packages containing malware/malicious links have been uploaded to the PyPI (Python Package Index) repository. Many of these have names which are slight misspellings of the names of other packages. The intention is to cause an installation of one of these packages if the package name is mistyped…