Skip to content

Logging in/Resources

Login takes place via the Jupyter URL for the individual HPC service:

anmeldung.gif

For login, your username, service password and a 2-factor authentication are required.

You will first find yourself on a landing page that also gives more information about the currently installed software versions.

By pressing the login button you will start the login process. Select the organization (e.g. KIT) that has granted you access to the HPC system and press Continue. In the Login section that appears, enter your username and password (not the HPC service password).

After pressing the Login button you will be redirected to the Second Factor query page. Enter the One-Time Password and click on Validate. Now you will be redirected to the JupyterHub page. Click on Enter JupyterHub to complete the login process.

Selection of the compute resources

Jupyter notebooks are executed in an interactive session on the compute nodes of the HPC clusters. Just like when accessing an interactive session with SSH, resource allocation is managed by the Slurm workload manager system. The selection of resources for Jupyter is possible using drop-down menus. Jupyter jobs can currently only run on a single node.

ressourcen.gif

Available resources for selection are:

  • Number of CPU cores
  • Number of GPUs
  • Runtime
  • Partition/Queue
  • Amount of main memory
  • Jupyter base module

In Normal Mode, the grayed-out fields contain reasonable presets depending on the number of required CPU cores or GPUs selected. The presects can be bypassed in Advanced Mode, where further options are available.

Advanced Mode can be activated by clicking on the checkbox of the same name. The following additional options then become available:

falsche_ressourcen.gif

  • Specification of a reservation
  • Selection of an account, if one is a member of several projects
  • LSDF mount option
  • BEEOND mount option

After the selection is made, the interactive job is started with the Spawn button. As when requesting interactive compute resources with the salloc command, waiting times may occur. These are usually the longer the larger the requested resources are. Even if the chosen resources are available immediately, the spawning process may take up to one minute.

Please note that in advanced mode, resource combinations can be selected that are impossible to be met. In this case, an error message will appear when the job is spawned.

The spawning timeout is currently set to 10 minutes. With a normal workload of the HPC facility, this time is usually sufficient to get interactive resources.

Prioritized access to computing resources on bwUnicluster 2.0

ampel.gif

The use of Jupyter requires the immediate availability of computing resources since the JupyterLab server is started within an interactive Slurm session. To improve the availability of GPUs for interactive supercomputing with Jupyter, automatic reservation for GPU (gpu_8) resources has been set up on bwUnicluster 2.0. It is active between 8am and 8pm every day. The reservation is automatically active if

  • no other reservation is set manually
  • advanced mode is disabled

To give you a better overview of the currently available resources, a status indicator has been implemented. It appears when selecting the number of required GPUs and shows whether a Jupyter job of the selected size can currently be started or not. Green means the selected GPU resources are available instantly. Yellow means only a single additonal job of the selected size can be started. Red means there are no GPU resources left that could satisfy the selected amount of resources.

If there are no more resources available within the reservation, you can try selecting a different amount of GPUs or activate Advanced Mode and select a different partition. Availability can be estimated using sinfo_t_idle, which is available when logging in via SSH.


Last update: May 26, 2021