Activity Feed

AI & ML interests

Large scale distributed AI model training, model parallelisation, low-level GPU acceleration, make GPUs go brrrrr

Recent Activity

julien-c  updated a Space about 17 hours ago
nanotron/README
lvwerra  new activity about 18 hours ago
nanotron/book:Update README.md
lvwerra  new activity about 18 hours ago
nanotron/book:Update README.md
View all activity

HF PRESS
book cover

The Ultra-Scale Playbook: Training LLMs on GPU Clusters

Essential reading for anyone scaling ML infrastructure
The knowledge on how to efficiently scale training to large GPU clusters has been well kept within a handful big industry labs. With this book, we set out to lift the veil and release a comprehensive resource on distributed training.
AUTHORS
Nouamane Tazi, Ferdinand Mom, Haolun Zhao, Phuc Nguyen, Mohamed Mekkouri, Leandro Werra, Thomas Wolf
AFFILIATION
Hugging Face
PUBLISHED
Jul 30, 2025

This book PDF is accessible with a PRO subscription.

The Nanotron team focus on sharing open knowledge and developping open-source libraries for efficient distributed training of large-scale AI models.

Some of its contributions are: