LRZ AI Systems
LRZ AI Systems / BayernKI
MAINTENANCE NOTICE
The AI Systems (including the BayernKI and MCML system segments) will undergo a maintenance procedure between May 19th and 21st, 2025. On these days, the system will not be available to users. Normal user operation is expected to resume during the course of Wednesday, May 21st.
"lrz-v100", "lrz-hpe-p100" and "lrz-cpu" SLURM partitions have been operating in degraded mode since 13.04.2024. No other partition will be affected by this ongoing worker nodes maintenance.
For further details see 99. AI Systems Announcements
NOTICE
This system is currently in pilot operation.
JOB SUBMISSION: --gres=gpu:X required
You must always indicate the --gres=gpu option when requesting a GPU resources allocation.
e.g., if you want to use 2 GPUs on a system, you must add --gres=gpu:2 when allocating resources
module, conda and pip not supported!
To define your software stack, make use of the Enroot container framework (see 4. Introduction to Enroot: The Software Stack Provider for the LRZ AI Systems).
Documentation
- 0. Getting Started
- 1. General Description and Resources
- 2. Storage on the LRZ AI Systems
- 3. Access and Getting Started
- 4. Introduction to Enroot: The Software Stack Provider for the LRZ AI Systems
- 5. Using NVIDIA NGC Containers on the LRZ AI Systems
- 6. Running Applications as Interactive Jobs on the LRZ AI Systems
- 7. Running Applications as Batch Jobs on the LRZ AI Systems
- 8. Multi-GPU Jobs on the LRZ AI Systems
- 9. Multi-Node Jobs on the LRZ AI Systems
- 10. Creating and Reusing a Custom Enroot Container Image
- 11. Interactive Web Servers on the LRZ AI Systems
- 12. Public Datasets and Containers on the LRZ AI Systems
- 98. AI Systems Reference
- 99. AI Systems Announcements