[2602.19945] DP-FedAdamW: An Efficient Optimizer for Differentially Private Federated Large Models

[2602.19945] DP-FedAdamW: An Efficient Optimizer for Differentially Private Federated Large Models

arXiv - AI 3 min read Article

Summary

The paper introduces DP-FedAdamW, a novel optimizer designed for differentially private federated learning, addressing key challenges in convergence efficiency and robustness.

Why It Matters

As federated learning becomes increasingly important for privacy-preserving machine learning, optimizing performance while maintaining data privacy is crucial. DP-FedAdamW offers a solution that enhances training efficiency and reduces client drift, making it significant for both researchers and practitioners in AI.

Key Takeaways

  • DP-FedAdamW stabilizes second-moment variance and removes bias in differentially private federated learning.
  • The optimizer achieves a linearly accelerated convergence rate without heterogeneity assumptions.
  • Empirical results show DP-FedAdamW outperforms existing methods by a notable margin on benchmark datasets.

Computer Science > Machine Learning arXiv:2602.19945 (cs) [Submitted on 23 Feb 2026] Title:DP-FedAdamW: An Efficient Optimizer for Differentially Private Federated Large Models Authors:Jin Liu, Yinbin Miao, Ning Xi, Junkang Liu View a PDF of the paper titled DP-FedAdamW: An Efficient Optimizer for Differentially Private Federated Large Models, by Jin Liu and 3 other authors View PDF HTML (experimental) Abstract:Balancing convergence efficiency and robustness under Differential Privacy (DP) is a central challenge in Federated Learning (FL). While AdamW accelerates training and fine-tuning in large-scale models, we find that directly applying it to Differentially Private FL (DPFL) suffers from three major issues: (i) data heterogeneity and privacy noise jointly amplify the variance of second-moment estimator, (ii) DP perturbations bias the second-moment estimator, and (iii) DP amplify AdamW sensitivity to local overfitting, worsening client drift. We propose DP-FedAdamW, the first AdamW-based optimizer for DPFL. It restores AdamW under DP by stabilizing second-moment variance, removing DP-induced bias, and aligning local updates to the global descent to curb client drift. Theoretically, we establish an unbiased second-moment estimator and prove a linearly accelerated convergence rate without any heterogeneity assumption, while providing tighter $(\varepsilon,\delta)$-DP guarantees. Our empirical results demonstrate the effectiveness of DP-FedAdamW across language and vision ...

Related Articles

Machine Learning

[D] ICML reviewer making up false claim in acknowledgement, what to do?

In a rebuttal acknowledgement we received, the reviewer made up a claim that our method performs worse than baselines with some hyperpara...

Reddit - Machine Learning · 1 min ·
UMKC Announces New Master of Science in Artificial Intelligence
Ai Infrastructure

UMKC Announces New Master of Science in Artificial Intelligence

UMKC announces a new Master of Science in Artificial Intelligence program aimed at addressing workforce demand for AI expertise, set to l...

AI News - General · 4 min ·
Machine Learning

[D] Budget Machine Learning Hardware

Looking to get into machine learning and found this video on a piece of hardware for less than £500. Is it really possible to teach auton...

Reddit - Machine Learning · 1 min ·
Machine Learning

Your prompts aren’t the problem — something else is

I keep seeing people focus heavily on prompt optimization. But in practice, a lot of failures I’ve observed don’t come from the prompt it...

Reddit - Artificial Intelligence · 1 min ·
More in Machine Learning: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime