Skip to content

Change Log

Updates to ScienceCloud and ScienceCluster are available on this page. For the Supercomputer service, please see the Alps (Eiger) User Guide.

2024

November

2024-11-06

  • ScienceCluster:
    • Users can get the updated list of modules and avoid module load errors by removing the lmod cache. We therefore suggest to run the following command:
      rm .cache/lmod/*
      
    • New released modules: VS Code 1.93 (code-server 4.93.1), jupyter 2024-10-22, RStudio Server with R 4.4.2, mamba 24.9.0-0, rclone 1.68.1, singularityce 4.2.1, boost 1.86.0, perl 5.40.0, cmake 3.30.5, gsl 2.8, python 3.13.0, nvhpc 24.9, cuda 12.6.2, nccl 2.22.3-1, cudnn 9.5.1.17-12, gcc 14.2.0, openmpi 5.0.5, hdf5 1.14.5, intel-oneapi-compilers 2025.0.0, intel-oneapi-mkl 2025.0.0, intel-oneapi-tbb 2022.0.0, intel-oneapi-mpi 2021.14
    • Removed unused and old modules: Matlab 2021b, VS Code Server 3.10.2, anaconda3 2022.05, 2022.10, mamba 23.3.1-1, perl 5.34.1, rclone 1.59.1, singularity 3.10.2, 3.11.3, 4.0.2, 4.1.0, cmake 3.26.3
    • Upgrade: nvidia driver

August

2024-08-22

  • ScienceCluster:
    • (ScienceApps) Upgraded VSCode. Added support for interactive GPU sessions.

July

2024-07-03

  • ScienceCluster:
    • Upgraded Slurm to version 24.05
    • New released modules: cmake 3.29.6, cudnn 9.2.0.82-12 (with cuda 12.4.1), gcc 14.1.0 (with cuda 12.4.1), openmpi 5.0.3, fftw 3.3.10 and hdf5 1.14.3, both recompiled with openmpi 5.0.3

June

2024-06-05

  • ScienceCluster:

    • The default value of 1MB is now set for the SLURM memory parameters. This will prevent the whole node allocation when no memory parameters is specified while running sbatch/srun commands.

      Important: adjust your SLURM submission scripts so that one of the memory parameters is specified (--mem, --mem-per-cpu, --mem-per-gpu), otherwise the default value of 1MB will be used.

    • New released modules: python 3.12.3, anaconda3 2024.02-1, mamba 24.3.0-0, cuda 12.4.1, cudnn 8.9.7.29-12, jupyter 2024.06 (4.2.1), RStudio Server 2024.06 (R 4.4)

    • Removed unused and old modules: boost 1.78.0, boost 1.83.0, python 3.11.0, mamba 4.14.0-0, mamba 22.9.0-2, mamba 23.1.0-1, nvhpc 22.7, rclone 1.62.2
    • Hardware changes: replacement of the large memory nodes and several CPU nodes of medium memory added.

May

2024-05-08

  • ScienceCluster:
    • System upgrades for improved stability and security
    • singularityce 4.1.0 is now the default version

March

2024-03-06

  • ScienceCloud:
    • Access to the the cloud console (cloud.s3it.uzh.ch) has been restricted to UZH-only networks (wired, wifi, and VPN). The cloud console is now dropping external non-UZH connection attempts.

February

2024-02-07

  • ScienceCluster:
    • Users can now use the command "ls -lh" to print the content size for each folder/file: this allows to quickly find large folders in case of over-quota issues, including hidden ones (e.g. ".local").
    • Upgrade: nvidia driver
    • New released modules: cuda 12.3.0, openmpi 5.0.1, boost 1.84.0, mamba 23.11.0-0, rclone 1.65.1, singularityce 4.1.0

2023

December

2023-12-04

  • ScienceCluster:
    • Updates to modules: python 3.12, anaconda3 2023.09-0, mamba 23.3.1-1, nvhpc 23.9, singularityce 4.0.2, cuda 12.2.1, nccl 2.19.3-1, cudnn 8.9.5.30-12.2, cudnn 8.9.5.30-11.8, gcc 13.2.0, openmpi 4.1.6

2023-11-09

  • ScienceCluster:
    • Matlab R2023b module released
  • ScienceCloud:
    • Images updated; see here for the list of active images
      • MatLab and Gaussian users should use a "***Singularity..." image then follow these MatLab and/or Gaussian instructions

2023-11-01

  • ScienceCluster:
    • OnDemand (for ScienceApps) upgraded for improved stability and security
    • Upgraded kernel version and firmware on compute nodes for improved stability and security
    • Users can now take advantage of all GPU nodes via the lowprio partition
    • To fix long-standing issues, some changes will be made to the user environments. Because of that, we recommend performing the following actions.

      Important: this should be done when no jobs are running; pending jobs are fine.

      • 1 - Move $HOME/.local to your data directory and symlink it. But first, check if the destination already exists:
        stat /data/$USER/.local
        
      • 2 - If it shows "stat: cannot stat ..." then run:
        mv $HOME/.local /data/$USER
        ln -s /data/$USER/.local $HOME/.local
        
      • 3 - Otherwise, you'll need to merge these two directories. Since each user case would be unique, we are unable to provide specific instructions. Please contact Science IT if you have any specific questions.