Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.

HideElements
breadcrumbtrue
titletrue
spacelogotrue

<< Zurück zur Dokumentationsstartseite

Lrz box
Picture/images/lrz/Icon_HPC.png
Heading1High Performance Computing

Forgot your Password? click here
Add new user (only for SuperMUC-NG)?
click here

Add new IP(only for SuperMUC-NG)?
click here
How to write good LRZ Service Requests? click here


System Status (see also:

 

Access and Overview of HPC Systems)

Status
colourGreen
= fully operational
Status
colourYellow
= operational but experiencing problems with restrictions (see messages below)
Status
colourRed
= not available



SuperMUC Phase 2

(will be switched off  , all data will be deleted)

Status
colourYellow
titleEND OF LIFE

login: hw

Höchstleistungsrechner (SuperMUC-NG)

System: 


login nodes: skx.supermuc.lrz.de

Status
colour

Green

Red
title

UP

down

File Systems:

Status
colourYellow
titleEND OF LIFE

Queues: micro, general, test, big

Status
colourGreen
titleUP

Detailed node status: 

Status
colourGreen
titleUP

login: skx
archive nodes: skx-arch.supermuc.lrz.de

Status
colour

Green

Red
title

UP

down

File Systems

: 


HOME

:


WORK

:


SCRATCH

:


DSS

:


DSA


Status
colourGreen
titleup

Status
colourGreen
title
UP
up

Status
colourGreen
title
UP
up

Status
colourGreen
title
UP
up
Status
colourGreen
titleup

Partitions/Queues: 
micro,

fat,

general, large

fat, test


Status
colourRed
titledown

Status
colour

Green

Red
title

UP

down

 Globus Online File Transfer: 

Status
colourGreen
title

UP

up

Detailed node status


Details:

Submit an Incident Ticket for the SuperMUC-NG

 

Add new user? click here

Add new IP? click here



Linux Cluster

 
CoolMUC-2
login:

lxlogin(
5
1,2,
6
3,
7
4).lrz.de

Status
colour

Redlxlogin(8, 10).lrz.de

Green
titleup

serial partitions: serial

Status
colourGreen
title

MAINTENANCE

up

parallel partitions cm2_(std,large)

Status
colour

Red

Green
title

MAINTENANCEPartitions/Queue:mpp2_batch

up

cluster cm2_tiny

Status
colour

Redmpp2

Green
title

MAINTENANCE

up

interactive partition: cm2_inter

Status
colour

Red

Green
title

MAINTENANCE

up

serial


c2pap

Status
colour

Red

Green
title

MAINTENANCE

up

CoolMUC-3

lxlogin(8,9).lrz.de

parallel partition: mpp3_batch

kcs

interactive partition: mpp3_inter


Status
colourGreen
titleup

Status
colour

Red

Green
title

MAINTENANCE

up

Status
colour

Red

Green
title

MAINTENANCEmpp3_inter

up

teramem, kcs

teramem_inter


ivymuc

kcs

Status
colourGreen
titleUP

Status
colourGreen
titleup

File Systems

HOME
SCRATCH
DSS
DSA


Status
colourGreen
titleup

Status
colour

Red

Green
title

MAINTENANCE

up

Status
colour

Red

Green
title

MAINTENANCE

up

Status
colour

Red

Green
title

MAINTENANCE

up

Detailed node status

click here


Detailed queue status


Details:

Submit an Incident Ticket for the Linux Cluster



Compute Cloud and
other HPC Systems

 

Compute Cloud: (https://cc.lrz.de)

Status
colourGreen
titleUP

GPU Cloud (

detailed status and free slots: https://

datalab.srv

cc.lrz.de

)

/lrz

Status
colourGreen
title

UP

DGX-1

Status
colourGreen
titleup

DGX-1v

up

LRZ AI Systems

Status
colourGreen
titleUP

RStudio Server

(https://www.rstudio.lrz.de)

Status
colour

Green

Red
title

UP

End of LIfe

Details:

Dokumentation
RStudio Server (LRZ Service)
  • Consulting for HPC and BigData Services at LRZ
  • Submit an Incident Ticket for the Compute Cloud

    Submit an Incident Ticket for RStudio Server



    Messages

    of the Day

    see also: Aktuelle LRZ-Informationen / News from LRZ



    Messages for SuperMUC

     Change of Access Policy for the tape archive

    Due to changed technical specifications for the IBM Spectrum Protect software, we have to change the access policy for the tape archive on SuperMUC-NG.
    This will also affect data from SuperMUC, which have already put into the tape archive.

    • Permissions to access the data will now be granted to all users of a project i.e., all users in a project group can retrieve data from other users in this project group.
    • The previous policy was that only the users who wrote the data into the archive could access it.
    • If your project is ‘pr12ab’, you can see the members of this group by
      getent group pr12ab-d
    • You have to add the project in the dsmc commands  i.e.
      dsmc q ar “/gpfs/work/p12ab/us12ab5/*“ –se=p12ab
    • Please note the difference between the project (“pr12ab”) and the permission group for data (”pr12ab-d”)

    See also: Backup and Archive on SuperMUC-NG

     Deletion of data on SuperMUC Phase 2.

    Required on your part: Data Migration from SuperMUC to SuperMUC-NG!

     skx-arch.supermuc.de (Node for archiving) will not be available before January, 2020

    Messages for Linux Cluster

     SCRATCH/GPFS unavailable on CoolMUC-3

    To avoid file system crashes, we have decided to unmount SCRATCH from all systems associated with CoolMUC-3 for now. We expect to revert this measure again at the next scheduled maintenance .

     Maintenance

    : HOME directory path will change

    You will find all HOME data in the new DSS HOME area, data migration will be performed by LRZ (unless you are specifically notified that you need to perform HOME data migration yourself). For emergency recoveries, the legacy NAS area (pointed to by the HOME_LEGACY variable) will remain available in read-only mode until the end of the year.

    Following action is required on your part:  Make necessary adjustments to (job) scripts and configuration data to account for the changes in path names. LRZ strongly recommends using relative path names because this minimizes the required work. 

    Examples: 

     End of service for NAS systems

    NAS paths will be taken offline at the beginning of January, 2020. Please contact the Service Desk for outstanding data migration issues.

    NAS PROJECT path is mounted read-only

    between and  

    DSS PROJECT now available on HPC systems:

    Following action is required on your part:

    Migrate data from the legacy NAS area (pointed to by the PROJECT_LEGACY variable) to the new DSS area. LRZ strongly advises to get rid of unneeded data sets, and/or archive data sets to tape.

    Step-by-step procedure for migration:

    1. On any cluster login node, issue the command
      dssusrinfo all
      This will list paths to accessible containers, as well as quota information etc.
    2. Edit your shell profile and set the PROJECT and/or WORK variable to a
      suitable path value based on the above output, typically one of the DSS
      paths with your account name appended to it.
    3. Use the cp or rsync or tar command to migrate your data from
      PROJECT_LEGACY to the new storage area.
    4. If your scripts use absolute path names instead of the PROJECT or
      WORK variable, they need appropriate updates

    https://www.lrz.de/aktuell/ali00788.html

    User e-Mail notification of DSS PROJECT link

    Following actions are required on your part:

    1. Confirm the e-Mail invitation to validate your access,
    2. After the Linux Cluster maintenance (see below), store path information in an environment variable on the Cluster (e.g. by setting the PROJECT variable in ~/.bashrc).

    One this is done, migrating data from NAS PROJECT to DSS PROJECT can start

    Please read the change description on how to handle the significant changes to the Linux cluster configuration performed end of September, 2019

    Messages for Cloud and other HPC Systems

                                            

    -NG

    Please read https://www.lrz.de/aktuell/ali00938.html for details on a scheduled maintenance starting May 16.

    On Monday, April 4, 2022, a new version of the spack-based development and application software stack will be rolled out.

    The new spack version will be loaded as default starting April 11, 2022

    After that date, you will be still able to switch to the previous spack stack with

    > module switch spack spack/21.1.1

    We strongly recommend recompiling self-built applications after the roll-out. See also https://doku.lrz.de/display/PUBLIC/Spack+Modules+Release+22.2.1 for details.

    Base core frequency of jobs has been set to 2.3GHz. Higher frequencies possible using EAR.

    The new hpcreport tool is now available to check job performance and accounting on SuperMUC-NG. Please check out

    https://doku.lrz.de/display/PUBLIC/HPC+Report

    https://www.lrz.de/aktuell/ali00923.html



    Messages for Linux Clusters

    There are 4 "new" Remote Visualization (RVS_2021) nodes available. The machines are in production mode. Nodes are operated under Ubuntu OS and NoMachine. Usage is limited to 2 hours and if you need a longer period of time, please file an LRZ Service Request. For more details please refer to the documentation.



    Messages for Cloud and other HPC Systems

    We have observed and addressed an issue with the LRZ AI Systems that concerned some running user jobs. As of now, newly started jobs should not be affected anymore. 
    The work on the LRZ AI Systems to address the recently observed stability issues has been concluded. All users are invited to continue their work. We closely monitor system operation and will provide additional updates if needed. Thank you for your patience and understanding.

    We have identified the likely root cause for the ongoing issues with the LRZ AI and MCML Systems following the latest maintenance downtime. We continue work towards a timely resolution and can currently not guarantee uninterrupted & stable system availability. For further details, please see LRZ AI Systems

    The LRZ AI and MCML Systems did undergo a maintenance procedure from April 25th to April 27th (both inclusive.) During this period, the system was not available to users. Normal user operation did resume on 2022-04-27 16:30.



    More Links

    Children Display