TSD Operational Log - Page 6
Hosts mounting /cluster may be experiencing NFS hangs at the moment. We're actively working on a solution.
We are investigating some reported login problems with data import and export. We will come back with an update once we have gathered more info.
Update 15:37: The problems have been resolved.
Currently, we are experiencing problems with managing groups of a TSD-project via TSD Selfervice, while logging in with ID-Porten (MinID, BankID, Buypass, Commfides). As a temporary workaround, please log in via TSD Credentials, to perform this task of managing groups in your TSD-project.
Update May 20: The problems have been resolved
Some projects experienced /cluster NFS hangs on April 25th between 19:00 and 19:45 and April 26th between 06:30 and 08:00.
We do not expect there to be any interruptions.
Som informert tidligere i ?r (ca. slutten av januar) skulle vi fra 1. mai ha innf?rt lisenskostnader for Windows i TSD. TSD rapporterer n? bruk av Microsoft-produkter i TSD til Microsoft p? m?nedsbasis, basert p? antall personer med faktisk tilgang. Grunnet noen sm? tekniske utfordringer har vi n? valgt ? utsette avregningen til 1. juni.
Innen 1. juni vil Prosjektleder i TSD, via selvbetjeningsportal, kunne styre hvem som skal ha tilgang til de ulike tjenestene ved ? melde folk inn og ut av grupper. Vi vil publisere fremgangsm?ten for ? styre inn-, og utmelding av prosjektets medlemmer p? denne lenken:
Login to TSD is currently unavailable.
We are working to solve the problem as quickly as possible.
Our apologies for the inconvenience.
--
The TSD Team
All RHEL6 ThinLinc (pxx-tl01-l) machines have now been shut down, as mentioned in the email sent in february. With a few exceptions.
A new RHEL8 Machine has also been made available to every project which can be accessed at https://view.tsd.usit.no
Read: /english/services/it/research/sensitive-data/use-tsd/login/index.html#toc8
If you for any reason need to access your RHEL6 Machine for a limited time, please contact us:/english/services/it/research/sensitive-data/contact/index.html
Update 20:00 April 27: a few submit and login hosts that mount /cluster are experiencing new NFS hangs. Some host have been rebooted.
There were NFS hangs on submit and login nodes that mount /cluster.
We are performing network maintenance on Thursday 29/4/2021.
We do not expect there to be any interruptions.
The cost command, used to query cpu quota usage on Colossus, is currently not working for projects without Sigma2 quota.
Update: the cost command now displays usage stats for Sigma2 quota, and will display NA and an info message for projects without Sigma2 quota.
Starting from April 1st., we will be introducing the following changes in the distribution of Colossus Quotas:
- We will reduce the Sigma2 pool of resources to 1536 cpu cores, with no gpu nodes. Only TSD-projects with cpu hour quota from Sigma2 can use this pool.
- We will move the removed resources from the Sigma2 pool to a dedicated resource, called “tsd”, consisting of 288 cpu cores on ordinary compute nodes, plus 128 cpu cores and 4 gpu cards on two gpu nodes.
- All TSD-projects can use the “tsd” resource, by submitting jobs using "--account=pNN_tsd" instead of "--account=pNN". Please check this document, for the complete procedure:
/english/services/it/research/sensitive-data/use-tsd/hpc/dedicated-resources.html - There will be a limit of 200,000 cpu hours on “tsd” resource, as it is limited. However, we may increase this limit in future.
Login through VMware was unavailable for some hours last evening.
Update 21:20: Issue resolved.
The TSD Team
IDPorten is having technical problems. When they are resolved everything will continue normally
We're experiencing NFS hangs on many Linux hosts mounting /cluster since 5:55 this morning.
Its also affecting /cluster on the Colossus compute nodes. The majority of compute nodes have been rebooted which may have affected running jobs.
Update 12:00: The submit hosts and Colossus are currently unavailable.
Update 14:00: The issue has been resolved, and we're rebooting the submit hosts now.
Due to an outage, login through VMware is currently unavailable.
You should however still be able to login through https://login.tl.tsd.usit.no if your project has a Linux-VM.
We are working on getting things back to normal as quickly as possible.
--
The TSD Team
The storage system for project storage (not Colossus) is having performance issues. This is causing instability in file import and export, and some slowness on virtual machines. We are debugging and fixing this.
Consent Portal from registration temporary due to service modification. This does not mean that consent is no longer acquired. The consents will be delivered to your project normally. The already registered forms will continue to be exposed to consenters on the external portal. We expect to resume form registration in couple of days.
We are working on solving an issue with Microsoft Office in TSD, giving this error:
?Microsoft office can't find your license for this application. A repair attempt was unsuccessful or was canceled. Microsoft office will now exit"
We will update the progress here, once the issue is resolved.
Update 14:30: Maintenance is complete, and submit hosts are now being rebooted.
Colossus will have downtime Thursday 21 January from 12:00-14:00 due to a third party issue.
Colossus and submit hosts will not be available during this time. Any pending jobs will automatically be rescheduled after the downtime.
This message will be updated once the maintenance is complete.
Unfortunately our login service is down and you will not be able to log in.
We are working on bringing everything back as quickly as we can, and will update this message as we move forward with solving the issue.
--
Best regards,
The TSD team
Update (21:45) - Maintenance is completed and submit hosts are being rebooted.
Update (16:00) - The hardware upgrade is taking longer than anticipated and it extended until further notice.
IBM is performing hardware replacement on the ESS storage on monday from 12:00-16:00.
Colossus and submit hosts will not be available during this time.
This message will be updated once the maintenance is complete.
We are experiencing issues with the Colossus storage system, and have reached out to the vendor for technical support, with the highest priority. Some projects' submit hosts may experience NFS hangs.
We're experiencing problems with the ESS storage, affecting /cluster NFS mounts and login to submit hosts and RHEL login nodes.
There might also be interruptions to HPC jobs.