HDPC 2026
The 1st Tutorial of "When the Error-Bounded Lossy Compression Meets Large-Scale AI Model Training and Fine-Tuning in Federated Environments"
July 13-16, 2026
Cleveland, OH, USA
Tell Me More

Introduction

Federated learning has emerged as an effective approach for scaling large-scale AI model training across geographically distributed data silos while preserving the privacy of training data.However, communication efficiency and privacy preservation remain two primary challenges in the practical deployment of federated learning systems, largely due to the substantial transmission of model gradients and parameters over public networks with limited bandwidth. Error-bounded lossy compression has proven to be an effective technique for compressing model parameters and gradients, thereby reducing communication overhead while simultaneously addressing privacy issues.

In this tutorial, we will provide an overview of the background and current research landscape in federated learning, error-bounded lossy compression, and differential privacy.

Topics

We will provide both the Lecture and Hands-On Exercises in this tutorial. The tutorial will cover basic and advanced topics of applying the error-bounded lossy compression for large-scale model training in federated learning environments. The tutorial is structured into 7 main sections.

Section 1

we provide a comprehensive introduction to errorbounded lossy compression, including its underlying concept, motivation, and representative use cases. We also present an in-depth overview of large-scale model training within federated learning environments, covering its fundamental principles, parameter aggregation strategies, existing bottlenecks, limitations of current solutions, and open frameworks. Finally, we offer a detailed introduction to differential privacy, discussing its core concept, mechanisms for data protection, and applications in federated learning.

Section 2

we present the motivation for employing errorbounded lossy compression to reduce communication overhead during large-scale model training in cross-silo federated learning systems. We further introduce an algorithm designed to estimate an appropriate error bound, thereby ensuring model accuracy while applying error-bounded lossy compression to model parameters. Additionally, we discuss resource utilization in large-scale model training within cross-silo federated learning environments, including CPUs, GPUs, and networks, and identify potential strategies for further enhancing system performance. Finally, we provide experimental results based on widely adopted benchmarks.

Section 3

we present the motivation for employing errorbounded lossy compression of model parameters to simultaneously achieve communication efficiency and fairness in large-scale model training within cross-silo federated learning environments. Additionally, we propose an algorithm for estimating an appropriate error bound based on the information entropy of the training data and model performance, thereby ensuring both communication efficiency and fairness. Furthermore, we provide a comprehensive theoretical analysis of the aggregation error introduced by parameter compression. Finally, we present an extensive experimental evaluation using widely used benchmarks.

Section 4

we will present the motivation for introducing differential privacy (DP) guarantees into the quantization process and review state-of-the-art DP-quantization approaches. We begin with a general guideline for injecting DP-guaranteed noise into the quantization process, followed by a taxonomy of existing DP-quantization methods and an overview of several representative techniques. We then introduce a unified analytical framework that we propose, which integrates standardized DP proofs and consistent utility analysis. Finally, we present key insights from our comprehensive reproduction and replication study of existing DP-quantization methods, and discuss their practical trade-offs, highlighting how these findings can inform the design of future DP-quantization mechanisms.

Section 5

we will introduce APPFL, a highly modular opensource framework designed to make privacy-preserving federated learning both practically deployable and easy to extend. It allows researchers and practitioners to freely mix and match FL algorithms, differential privacy mechanisms, communication protocols (MPI for HPC simulation and gRPC for real cross-institution deployment), models, and datasets in a true plug-and-play fashion. We will also introduce the core technical contribution about IIADMM, a new communication-efficient algorithm based on inexact ADMM that completely eliminates the transmission of dual variables from clients to server (unlike the earlier ICEADMM), thereby cutting communication volume roughly in half while matching or exceeding the accuracy of FedAvg and ICEADMM.

Section 6

we introduce FedDES, a super-fast discrete-event simulator built specifically to predict the real-world runtime of huge federated learning systems with hundreds of thousands of devices. Instead of painfully running actual training tasks, FedDES profiles a small real deployment once, extracts the key costs (training time, model upload/download, aggregation), and then turns the entire FL process into a handful of timed events. Using SimGrid’s accurate network and compute modeling plus a smart priority-queue scheduler, it simulates synchronous (FedAvg), asynchronous (FedAsync), and semi-asynchronous (FedCompass) strategies under realistic heterogeneity, stragglers, and network conditions

Section 7

we provide a practical guide on deploying APPFL, an open framework designed for large-scale model training in federated learning environments. Additionally, we present a case study demonstrating the process of training large-scale models using this framework. To further analyze system performance in large-scale training scenarios, we introduce FedDES, a simulator capable of modeling collaboration among a large number of nodes during model training. We detail the deployment process of FedDES and illustrate how it can be utilized to evaluate system performance.

Tentative Schedule

Introduction

Introduction of the Tutorial(5 min.) by Dr. Sheng Di.

Session 1

Background of Error Bounded Lossy Compression and Large-Scale Model Training in Federated Environments (45 min.), by Dr. Sheng Di, Dr. Zhaorui Zhang, and Prof. Xiaodong Yu.

  • The introduction of the error-bounded lossy compression: includes the motivation of lossy compression; what is the error-bounded lossy compression; the most popular errorbounded lossy compressors; their use-cases;
  • The introduction of the large-scale model training in federated learning environments: includes what is federated learning; the parameter aggregation strategies; the bottlenecks of such tasks; the limitations of existing solutions; and the open frameworks for federated learning.
  • The introduction of differential privacy (DP): includes what differential privacy is, how differential privacy protects the data, and how to apply DP in federated learning.

Session 2

Error-Bounded Lossy Compression for Communication Reduction for Large-Scale Model Training in Federated Environments ( 20 min.), by Dr. Zhaorui Zhang.

  • The motivation why we want to using error-bounded lossy compression to compress the parameters for large-scale model training in federated learning systems.
  • The algorithm to estimate an appropriate error bound that ensures the model accuracy while applying error-bounded lossy compression to model parameters.
  • The resource utilization (CPUs, GPUs, and network) analysis and system performance optimization.
  • Experiment results analysis about the error-bounded lossy parameter compression.

Session 3

Towards Fairness and Communication Efficiency for Large-Scale Model Training in Federated Environments based on Error-Bounded Lossy Parameter Compression ( 20 min.), by Dr. Zhaorui Zhang.

  • Motivation to use the error-bounded lossy compression to compress parameters to achieve the communication efficiency and fairness simultaneously.
  • The algorithm to estimate an appropriate error bound for error-bounded lossy compressors that can guarantee the model accuracy and fairness simultaneously.
  • Theoretical analysis for the aggregated error caused by the lossy parameter compression.
  • Experiment results analysis about the error-bounded lossy parameter compression.

BREAK(30 min)

Session 4

Differentially Private Quantization for PrivacyPreserving and Communication-Efficient Federated Training ( 20 min.), by Prof. Xiaodong Yu.

  • The motivation why differential privacy (DP) should be integrated into quantization, and an overview of state-of-the-art DP-quantization approaches.
  • Generic guidelines for injecting DP noise into the quantization pipeline and a structured categorization of existing DP-quantization methods.
  • Introduction of representative DP-quantization techniques and our unified framework with standardized DP proofs and consistent utility analysis.
  • Key findings from reproduction/replication studies and practical implications for designing next-generation DP-quantization mechanisms.

Session 5

APPFL: Open-Source Software Framework for Privacy-Preserving Federated Learning (20 min.), by Prof. Xiaoyi Lu.

  • We will introduce APPFL, a highly modular, open-source Python framework for privacy-preserving federated learning that allows plug-and-play combination of FL algorithms, differential privacy mechanisms, communication protocols (MPI and gRPC), models, and datasets.
  • We will introduce a new communication-efficient algorithm called IIADMM of APPFL, which eliminates the need to transfer dual variables from clients to server while achieving equal or better accuracy than FedAvg and ICEADMM.
  • We will introduce how the APPFL integrates Laplace-based output perturbation differential privacy and the experiment results of APPFL on multiple datasets that IIADMM consistently outperforms existing methods under strong privacy budgets (ε = 10-3).

Session 6

FedDES: Discrete Event Performance Simulation for Large-Scale Federated Learning ( 20 min.), by Prof. Xiaoyi Lu.

  • The motivation why we want to proposing FedDES.
  • The key design principle of the FedDES. How the FedDES simulates the large-scale federated learning.
  • What kinds of training strategies can FedDES simulate for federated learning?
  • The key results that FedDES achieved.

Session 7

Hands-On Exercises on Useful Federated Training Frameworks and Compression Tools. ( 25 min.), by Prof. Xiaoyi Lu and Dr. Zhaorui Zhang.

  • How to deploy APPFL for federated learning;
  • How to deploy FedDES, which is the simulator for federated learning and used for system performance analysis for largescale model training.
  • How to integrate the error-bounded lossy compressors (such as LibPressio) into federated learning frameworks for communication reduction and privacy preserving.

Conclusion(5 min.)

Give a brief summary and conclude the tutorial, by ALL.

Speakers

Picture of TBD

Sheng Di

Computer Scientist with Argonne National Laboratory, USA.

Picture of TBD

Xiaoyi Lu

Associate Professor at the University of California, Merced.

Picture of TBD

Xiaodong Yu

Assistant Professor at Stevens Institute of Technology.

Picture of TBD

Zhaorui Zhang

Research assistant professor at the Hong Kong Polytechnic University.

About

  • Theme, Vision, Mission

    Read more about our Theme, Vision, and Mission

    Learn More
  • Agenda

    Check out the agenda

    Learn More
  • Team

    Find out the team members

    Learn More
  • Venue

    Learn more about our location

    Learn More
  • FAQs

    Have a Question? Many answers can be found right here in our FAQs

    Learn More
  • Event Booklet

    View or Download our event booklet

    Learn More

Contact Us

We will do our best to answer your request as soon as possible. You may send us your request via Facebook Message