Castiel2 Multi-GPU AI Train the Trainer workshop
Prerequisites
You are familiar with AI frameworks.
Schedule
The course will run across 5 days with a morning and afternoon session each day. Times in CET:
📅 Dates: 30th January - 5th February, 2026
🕘 Time: 9:30 AM – 5:30 PM daily
🍴 Lunch Break: 1:00 PM – 2:00 PM
Day |
Session |
Topic |
1 |
Morning |
Introduction to GPU architectures and Access to HPC infrastructure used |
1 |
Afternoon |
Introduction to deep learning |
2 |
Morning |
Pytorch Distributed Data Parallel |
2 |
Afternoon |
Model parallelism with Pytorch |
3 |
Morning |
Pytorch Lightning |
3 |
Afternoon |
LLM, Finetuning & HuggingFace Accelerate, Deepspeed |
4 |
Morning |
Computer Vision |
4 |
Afternoon |
MLops |
5 |
Morning |
Ray + Retrieval Augmented Generation (RAG) |
5 |
Afternoon |
Hyperparameter tuning followed by Q&A |
The lesson materials
- 1M: Access to Leonardo
- 1A: Introduction to Deep Learning
- 2M: PyTorch Distributed Data Parallel
- 2A: Model parallelism with Pytorch
- 3M: PyTorch Lightning
- 3A: Fine-tuning neural networks
- 3A: Accelerate FSDP Fine-Tuning
- 4M: Computer Vision with CNN
- 4A: MLOps on HPC
- 5M: Introduction to Ray
- 5M: Retrieval Augmented Generation (RAG)
- 5A: Hyperparameter tuning
Reference
Who is the course for?
Scientists, engineers, and PhD students who need to scale AI models efficiently or optimize distributed training workflows.
About the course
Instructors and workshop helpers
Instructors
Alberto Bocchinfuso (CINECA)
Ashwin Mohanan (RISE / ENCCS)
Elena Paraschiv (RoNCC)
Gyula Ujlaki (NCC Hungary)
Oskar Taubert (CSC)
Riccardo Scheda (CINECA)
Workshop helpers
Tobias Haas, HLRS
Vlasta Petrovic, BSC / Castiel2
Sima Barzegar, BSC / Castiel2
Eleonora Bergamaschi, CINECA
… and many more
Course pages
See also
…
Credits
…