Composition Theorems and Privacy Budget Management
Applying a differential privacy mechanism, such as noise addition, in a single round of federated learning provides an (ϵ,δ) guarantee for that specific interaction. However, federated learning typically involves numerous communication rounds, sometimes hundreds or thousands. Each round potentially leaks a small amount of information. A critical question arises: how does the overall privacy loss accumulate across these multiple rounds? Simply stating the per-round guarantee is insufficient; we need to understand the total privacy cost over the entire training process. This is where composition theorems become essential.
Understanding Privacy Loss Accumulation
Imagine repeatedly querying a database (or in our case, repeatedly receiving noisy aggregated updates derived from client data) using DP mechanisms. Each query introduces some privacy loss. Composition theorems provide bounds on the total privacy loss after a sequence of such queries or mechanisms.
Basic Sequential Composition
The simplest way to analyze the combined effect is through basic sequential composition. It states that if you apply k mechanisms, where the i-th mechanism is (ϵi,δi)-differentially private, the sequence of all k mechanisms is (∑i=1kϵi,∑i=1kδi)-differentially private.
If all mechanisms are identical, providing (ϵ,δ)-DP each time, then after k rounds, the total privacy loss is bounded by (kϵ,kδ).
ϵtotal≤kϵroundδtotal≤kδround
While simple and intuitive, this linear accumulation often results in a loose upper bound. In a typical FL scenario with many rounds (k large), kϵ can quickly become very large, suggesting a weak overall privacy guarantee unless the per-round ϵround is made extremely small. Making ϵround too small, however, often requires adding excessive noise, significantly degrading model utility. For practical FL systems needing hundreds of rounds, basic composition is often too pessimistic.
Advanced Composition
Fortunately, tighter bounds exist under the framework of advanced composition. These theorems leverage the fact that the privacy losses don't simply add up linearly in the worst way possible, especially concerning the ϵ parameter.
A standard result in advanced composition (derived from analyzing the moments accountant or using techniques like Rényi Differential Privacy) shows that for a sequence of k mechanisms, each satisfying (ϵ,δ)-DP, the combined mechanism satisfies (ϵ′,δ′)-DP for any δ′′>0, where:
ϵ′≈2kln(1/δ′′)ϵ+kϵ(eϵ−1)δ′=kδ+δ′′
Let's break down the expression for ϵ′:
The dominant term for small ϵ is often the first one: 2kln(1/δ′′)ϵ. Notice the k dependency. This means the total ϵ′ grows roughly sublinearly with the number of rounds k, rather than linearly as in basic composition.
The second term kϵ(eϵ−1) captures higher-order effects. For small ϵ, eϵ−1≈ϵ, so this term behaves like kϵ2.
δ′′ is an additional parameter representing a small probability that the privacy guarantee fails, in addition to the accumulated base δ values (kδ). It's typically chosen to be small, for example, smaller than 1/N, where N is the dataset size.
The implication is significant: advanced composition allows for a much more favorable accumulation of privacy loss, especially for the ϵ parameter. We can afford a larger per-round ϵround for the same total budget ϵtotal compared to basic composition, leading to potentially better model utility.
Comparison of total privacy loss (ϵ) accumulation over k rounds using basic (linear) and advanced (sublinear) composition, assuming ϵround=0.1, δround=10−5, and δ′′=10−6 for the advanced calculation. Advanced composition yields significantly tighter bounds for larger k.
Modern privacy accounting methods, often based on Rényi Differential Privacy (RDP), provide frameworks for precisely tracking privacy loss under composition. RDP uses a different function (the RDP curve) to measure privacy loss, which composes very naturally. Tools like Google's dp-accounting library implement these techniques, allowing developers to accurately compute the (ϵ,δ)-DP guarantee for complex algorithms like DP-FedAvg run over multiple rounds.
Managing the Privacy Budget
The total acceptable privacy loss for the entire federated learning process is known as the privacy budget, typically denoted as (ϵtotal,δtotal). This budget represents the maximum cumulative privacy leakage we are willing to tolerate. Choosing appropriate values for ϵtotal and δtotal depends heavily on the sensitivity of the data, the application context, and any regulatory requirements (like GDPR or HIPAA). Common practice often aims for ϵtotal in the single digits (e.g., 1 to 10) and δtotal to be very small (e.g., less than 1/Nclients, where Nclients is the number of clients).
Once a total budget (ϵtotal,δtotal) is set for, say, T communication rounds, the task becomes budget allocation: deciding how much privacy loss (ϵround,δround) is permissible in each round such that the total loss remains within the budget according to composition theorems.
Several strategies exist:
Uniform Allocation: The simplest approach is to divide the budget roughly equally across rounds. Using the advanced composition intuition (ϵ∝k), one might set ϵround≈ϵtotal/T and δround=δtotal/T. This requires determining the total number of rounds T in advance.
Non-Uniform Allocation: It might be beneficial to spend more budget (allow larger ϵround) in the initial rounds when model updates are larger and potentially more informative for convergence, and less budget (smaller ϵround) in later rounds during fine-tuning. This requires careful planning and analysis.
Adaptive Allocation: The per-round budget could be adjusted dynamically based on the training progress, convergence speed, or other runtime factors. This is more complex but potentially optimizes the privacy-utility trade-off more effectively.
Practical Notes
Accounting Tools: Manually applying composition theorems can be complex and error-prone, especially with advanced techniques like RDP. Using dedicated privacy accounting libraries is highly recommended. These libraries take the parameters of the DP mechanism used in each round (e.g., noise multiplier σ, clipping norm C, sampling probability q in DP-SGD/DP-FedAvg) and the number of rounds, and compute the resulting (ϵ,δ).
Parameter Tuning: There's a direct relationship between the noise added (e.g., noise standard deviation σ relative to the clipping norm C), the per-round privacy loss ϵround, and model utility. Lower noise generally means better utility but higher ϵround, consuming the budget faster. Higher noise provides better per-round privacy (lower ϵround) but might hurt convergence or final accuracy. Tuning these parameters (σ, C, T, batch size, learning rate) while respecting the total privacy budget (ϵtotal,δtotal) is a core challenge in implementing DP-FL systems.
Budget Depletion: Once the pre-defined privacy budget is exhausted after T rounds, the training process must stop releasing differentially private outputs to maintain the guarantee.
Understanding and correctly applying composition theorems is fundamental for building federated learning systems with meaningful, quantifiable privacy guarantees over the entire training duration. Advanced composition and specialized accounting tools are indispensable for achieving practical privacy-utility trade-offs in multi-round federated processes.
Was this section helpful?
The Algorithmic Foundations of Differential Privacy, Cynthia Dwork and Aaron Roth, 2014Foundations and Trends® in Theoretical Computer Science, Vol. 9 (Now Publishers)DOI: 10.1561/0400000042 - A fundamental textbook establishing the principles of differential privacy, including basic and advanced composition theorems.
Deep Learning with Differential Privacy, Martín Abadi, Andy Chu, Ian Goodfellow, H. Brendan McMahan, Ilya Mironov, Kunal Talwar, Li Zhang, 2016Proceedings of the 2016 ACM SIGSAC Conference on Computer and Communications Security (Association for Computing Machinery)DOI: 10.1145/2976749.2978318 - Introduces the moments accountant for tighter privacy composition bounds in deep learning, a method relevant for multi-round federated learning.
Renyi Differential Privacy, Ilya Mironov, Kunal Talwar, Li Zhang, 2017Proceedings of the 36th ACM SIGMOD-SIGACT-SIGART Symposium on Principles of Database Systems (ACM)DOI: 10.1145/3035918.3035925 - Proposes Rényi Differential Privacy (RDP) as an alternative privacy measure that simplifies privacy accounting under composition.
Google's Differential Privacy Library, Google, 2024 - An open-source library that provides practical tools for differentially private computations, including advanced privacy accounting, as referenced in the section.