Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.

HideElements
breadcrumbtrue
titletrue
spacelogotrue

<< Zurück zur Dokumentationsstartseite

Lrz box
Picture/images/lrz/Icon_HPC.png
Heading1High Performance Computing

Forgot your Password? click here
Add new user (only for SuperMUC-NG)?
click here

Add new IP(only for SuperMUC-NG)?
click here
How to write good LRZ Service Requests? click here


System Status (see also:

 

Access and Overview of HPC Systems)

Status
colourGreen
= fully operational
Status
colourYellow
= operational but experiencing problems with restrictions (see messages below)
Status
colourRed
= not available



SuperMUC Phase 2

(only available until end of 2019)

Höchstleistungsrechner (SuperMUC-NG)

System: 

Status
colour

Yellow

Green
title

END OF LIFE

up

login nodes:
hw
skx.supermuc.lrz.de

Status
colourGreen
title

UP
Status
colourYellow
titleEND OF LIFE
Status
colourYellow
titleEND OF LIFE
Status
colourGreen
titleUP
Queues: micro, general, test, big

Status
colourGreen
titleUP

Detailed node status: 

Status
colourGreen
titleUP

login: skx

up

File Systems:
HOME:
WORK:
SCRATCH:

Important: Please migrate data to SuperMUC-NG asap! 

archive nodes: skx-arch.supermuc.lrz.de

Status
colour

Green

Yellow
title

UP

up

File Systems

: 


HOME

:


WORK

:


SCRATCH

:


DSS

:


DSA


Status
colourGreen
titleup

Status
colourYellow
titleUP

Status
colour
Green
Yellow
titleUP

Status
colourGreen
title
UP
up
Status
colourGreen
titleup

Partitions/Queues: 
micro

, fat

, general, large

fat, test


Status
colourGreen
title

UP

up

Status
colourGreen
titleup

 Globus Online File Transfer: 

Status
colourGreen
title

UP

up

Detailed node status


Details:

Submit an Incident Ticket for the SuperMUC-NG

 

Add new user? click here

Add new IP? click here



Linux Cluster

 
CoolMUC-2
login:

lxlogin(
5
1,2,
6
3,
7
4).lrz.de

Status
colourGreen
titleup

serial partitions: serial

Status
colourGreen
title

UPlxlogin(8, 10).lrz.dempp2_batch

up

parallel partitions cm2_(std,large)

Status
colourGreen
titleUP

Partitions/Queue:
cluster cm2_tiny

Status
colourGreen
titleUP

mpp2
interactive partition: cm2_inter

Status
colourGreen
title

UP

up

serial
c2pap

Status
colourGreen
title

UP

mpp3_batch

kcs

up

CoolMUC-3

lxlogin(8,9).lrz.de

parallel partition: mpp3_batch

interactive partition: mpp3_inter


Status
colourGreen
titleup

Status
colour

Yellow

Green
title

w/o SCRATCH

up

Status
colour

Yellow

Green
title

w/o SCRATCHmpp3_inter

up

teramem, kcs

teramem_inter


ivymuc

kcs

Status
colourGreen
titleup

Status
colour

Yellow

Green
title

w/o SCRATCH

up

File Systems

HOME
SCRATCH
DSS
DSA


Status
colourGreen
title

UP

up

Status
colourGreen
titleup

Status
colourGreen
titleup

Status
colourGreen
title

UP

up

Detailed node status

click here


Detailed queue status


Details:

 

Submit an Incident Ticket for the Linux Cluster



Compute Cloud and
other HPC Systems

 

Compute Cloud: (https://cc.lrz.de)

Status
colourGreen
titleUP

Open Nebula Coud (

detailed status and free slots: https://

www

cc.

cloud

lrz.

mwn.

de

)

Status
colourYellow
titleend of life

GPU Cloud (https:/

/

datalab.srv.

lrz

.de) StatuscolourGreentitleUP

DGX-1

Status
colourGreen
titleup

DGX-1v
LRZ AI Systems

Status
colourGreen
titleUP

RStudio Server

(https://www.rstudio.lrz.de)

Status
colour

Green

Red
title

UP

End of LIfe

Details:

Dokumentation
RStudio Server (LRZ Service)
Consulting for HPC and BigData Services at LRZ

Submit an Incident Ticket for the Compute Cloud

Messages of the Day


Messages for SuperMUC

 Scheduled Maintenance:

November 27, 2019, at 8:00 am

Important: Please speedily migrate data to SuperMUC-NG! skx-arch.supermuc.de is not yet available before early January, 2020Messages for Linux Cluster

 SCRATCH/GPFS on CoolMUC-2

Today (from morning until early afternoon), the SCRATCH file system was not available. Within that timeframe, jobs may have crashed. SCRATCH file system was made available again.

Short lasting outages of SCRATCH have been observed today on CoolMUC-2 and hosted cluster systems as well.

 SCRATCH unavailable on CoolMUC-3

To avoid file system crashes, we have decided to unmount SCRATCH from all systems associated with CoolMUC-3 for now. We expect to revert this measure again at the next scheduled maintenance .

 End of service for NAS systems

NAS paths will be taken offline at the beginning of January, 2020. Please contact the Service Desk for outstanding data migration issues.

 Maintenance

November 27, 2019, at 8:00 am
-NG

Archive nodes update
The  hardware maintenance has been rescheduled. Please expect a short downtime of the archive and backup servers (IBM spectrum protect) from 09:00 to 11:00 o’clock on Monday 24.01.22 .

 The new ANSYS Software Release, Version 2022.R1 has been installed and provided on SuperMUC-NG. For details please refer to the correspondig announcement:
https://www.lrz.de/aktuell/ali00931.html

The Energy Aware Runtime (EAR) has been reactivated. Please be aware that this may have an impact on job processing times.

Please note that WORK/SRATCH on SuperMUC-NG exhibit currently possible performance degradation under heavy I/O load. Take this into account when planning your job runtimes.

The new hpcreport tool is now available to check job performance and accounting on SuperMUC-NG. Please check out

https://doku.lrz.de/display/PUBLIC/HPC+Report

https://www.lrz.de/aktuell/

ali00807

ali00923.html

: HOME directory path will change

You will find all HOME data in the new DSS HOME area, data migration will be performed by LRZ (unless you are specifically notified that you need to perform HOME data migration yourself). For emergency recoveries, the legacy NAS area (pointed to by the HOME_LEGACY variable) will remain available in read-only mode until the end of the year.

Following action is required on your part:  Make necessary adjustments to (job) scripts and configuration data to account for the changes in path names. LRZ strongly recommends using relative path names because this minimizes the required work. 

Examples: 

NAS PROJECT path is mounted read-only

between and  

DSS PROJECT now available on HPC systems:

Following action is required on your part:

Migrate data from the legacy NAS area (pointed to by the PROJECT_LEGACY variable) to the new DSS area. LRZ strongly advises to get rid of unneeded data sets, and/or archive data sets to tape.

Step-by-step procedure for migration:



  • On any cluster login node, issue the command
    dssusrinfo all
    This will list paths to accessible containers, as well as quota information etc.
  • Edit your shell profile and set the PROJECT and/or WORK variable to a
    suitable path value based on the above output, typically one of the DSS
    paths with your account name appended to it.
  • Use the cp or rsync or tar command to migrate your data from
    PROJECT_LEGACY to the new storage area.
  • If your scripts use absolute path names instead of the PROJECT or
    WORK variable, they need appropriate updates
    Messages for Linux Clusters

    SCRATCH is now fully online again. While we expect older data that were temporarily inaccessible to be fully available again, data that were created in the last few days before the problems started might be corrupt and need to be renewed from tape archive (if one exists) or recreated. 
    There will be a reboot of one servers tomorrow, which however should not impact overall system operation.

    The new ANSYS Software Release, Version 2022.R1 has been installed and provided on the LRZ Linux Cluster systems (CM2, CM3 and RVS systems). For details please refer to the correspondig announcement:
    https://www.lrz.de/aktuell/ali00931.html

    The new release of Abaqus, Version 2022 (Dassault Systems Software) has been installed on both Linux Clusters CoolMUC-2 / CoolMUC-3 as well as on the RVS systems. The Abaqus documentation has been updated.

    The new release of SimCenter StarCCM+, Version 2021.3.1 (Siemens PLM Software) has been installed and provided on the LRZ HPC systems (CM2, CM3, SNG and RVS systems). For details please see the correspondig announcement:
    https://www.lrz.de/aktuell/

    ali00788Messages for Cloud and other HPC systems

    The OpenNebula Compute Cloud was decommissioned on

                                            

    User e-Mail notification of DSS PROJECT link

    Following actions are required on your part:

    1. Confirm the e-Mail invitation to validate your access,
    2. After the Linux Cluster maintenance (see below), store path information in an environment variable on the Cluster (e.g. by setting the PROJECT variable in ~/.bashrc).

    One this is done, migrating data from NAS PROJECT to DSS PROJECT can start

    Please read the change description on how to handle the significant changes to the Linux cluster configuration performed end of September, 2019

    There are 4 new Remote Visualization (RVS_2021) nodes available in a friendly user testing period. Nodes are operated under Ubuntu OS and NoMachine. For more details please refer to the documentation.



    Messages for Cloud and other HPC Systems

    The LRZ AI and MCML Systems are back in operation as the maintenance procedure planned from January 7thto January 11th is completed.

    The RStudio Server service at LRZ was decommissioned. For a replacement offering please see Interactive Web Servers on the LRZ AI Systems and, more generally, LRZ AI Systems.



    More Links

    Children Display