Loading Now

Differential Privacy: From Convergent Federated Learning to Safeguarding LLMs and Beyond

Latest 27 papers on differential privacy: May. 16, 2026

The quest for powerful AI models often clashes with the fundamental need for data privacy. Differential Privacy (DP) stands as a beacon in this challenge, offering mathematical guarantees for data protection. Recent breakthroughs in DP are not just refining existing techniques but are fundamentally reshaping how we approach privacy in complex AI/ML systems, from the intricacies of federated learning to the rapidly evolving landscape of Large Language Models (LLMs) and beyond.

The Big Ideas & Core Innovations

One of the most exciting developments challenges a long-held assumption in federated learning (FL). Traditionally, it was believed that privacy budgets in FL-DP would indefinitely increase with more training rounds. However, a groundbreaking analysis by Yan Sun et al. (University of Sydney, Nanyang Technological University, Sun Yat-sen University) in their paper, “Convergent Differential Privacy Analysis for General Federated Learning”, reveals that properly designed local regularization terms in methods like Noisy-FedProx can achieve convergent privacy bounds. This means privacy can stabilize rather than perpetually degrade, offering a win-win for both optimization and privacy in FL-DP by using a novel shifted interpolation technique.

While FL-DP makes strides, other challenges emerge. **Xiao Tian et al. (National University of Singapore, A*STAR), in “INO-SGD: Addressing Utility Imbalance under Individualized Differential Privacy”, pinpoint a critical utility imbalance in Individualized DP (IDP). They show that stronger privacy requirements for certain data owners can lead to their data being underrepresented, causing poor model performance for those groups. Their proposed INO-SGD algorithm** strategically down-weights less important gradients based on loss order, thereby improving utility for more private owners without sacrificing overall model performance. This directly addresses the “minority initial drop” and “biased optimization objective” phenomena inherent in IDP.

Privacy in LLMs is another hotbed of research. Zeyuan Chen et al. (CISPA Helmholtz Center for Information Security) unveil the “Pop Quiz Attack: Black-box Membership Inference Attacks Against Large Language Models”, a potent black-box membership inference attack that exposes training data memorization in LLMs, achieving an average ROC_AUC of 0.873. This highlights the persistent vulnerability of LLMs, even against existing defenses including DP, and underscores that larger models are often more vulnerable. Complementing this, Eduardo Tenorio et al. (University of Arkansas) in “How Does Differential Privacy Affect Social Bias in LLMs? A Systematic Evaluation”, demonstrate that while DP can reduce stereotypical bias in logit-level sentence scoring, this improvement doesn’t reliably transfer to output-level bias in tasks like text completion or question answering. This points to a crucial disconnect and the need for multi-paradigm evaluation.

For spatial data, Kennedy Edemacu et al. (The City University of New York, Northern Michigan University, Sangmyung University) introduce “Privacy Without Losing Place: A Paradigm for Private Retrieval in Spatial RAGs”. Their PAS (Privacy Anchor Substitution) mechanism enables location privacy in spatial Retrieval-Augmented Generation (RAG) systems by encoding locations using relative anchor triplets instead of exact coordinates, achieving significant adversarial localization error while retaining retrieval performance. This paradigm shift means systems don’t always need exact coordinates, opening new avenues for location-aware, privacy-preserving AI.

Beyond these, advancements are being made at the mechanism level. Daniel Cortild and Coralia Cartis (University of Oxford), in “Quadratic Objective Perturbation: Curvature-Based Differential Privacy”, introduce Quadratic Objective Perturbation (QOP), a novel DP mechanism for empirical risk minimization that uses random quadratic forms to control sensitivity through curvature, rather than bounded gradients. This innovation is particularly impactful in the interpolation regime, crucial for modern overparameterized models. In a similar vein, Mohammad Partohaghighi and Roummel Marcia (University of California, Merced) present “Deep Learning under Fractional-Order Differential Privacy”, introducing FO-DP-SGD. This method incorporates fractional-order memory directly into the private release mechanism (before noise addition), leading to improved test accuracy and privacy-utility performance over standard DP-SGD by selectively incorporating historical information.

Challenges in auditing AI systems are also being addressed. Florian A. D. Burnat and Brittany I. Davidson (University of Bath), in “Differentially Private Auditing Under Strategic Response”, formalize privacy-constrained auditing as a Stackelberg game. They show that naive DP auditing is suboptimal when developers can strategically respond, introducing the SPAD algorithm which uses hypergradients to achieve 5-44% reductions in under-detection gaps.

Under the Hood: Models, Datasets, & Benchmarks

These innovations are often built upon, and contribute to, a rich ecosystem of models, datasets, and benchmarks:

Impact & The Road Ahead

The landscape of differential privacy is rapidly expanding, showcasing its critical role in building trustworthy and ethical AI systems. The convergence of privacy guarantees in federated learning (Yan Sun et al.) means we can pursue long-term, privacy-preserving collaborations without an ever-increasing privacy deficit. Addressing utility imbalance in individualized DP (Xiao Tian et al.) highlights the importance of fairness within privacy mechanisms, ensuring that privacy doesn’t disproportionately disadvantage certain data owners. The revelations about LLM privacy (Zeyuan Chen et al., Eduardo Tenorio et al.) are a stark reminder that simply applying DP is often not enough; a deeper understanding of model behavior and multi-faceted evaluation is essential.

Beyond these, the shift towards curvature-based privacy (Cortild and Cartis), memory-enhanced DP-SGD (Partohaghighi and Marcia), and novel approaches to spatial privacy (Edemacu et al.) indicates a maturation of DP mechanisms. This allows for more nuanced and efficient privacy protection tailored to specific data types and model architectures. The work on strategic auditing (Burnat and Davidson) is crucial for real-world AI governance, acknowledging that actors will adapt to privacy mechanisms. Finally, the theoretical guarantees for private RL (Yi He and Xingyu Zhou) and efficient exact samplers (Markus de Medeiros et al.) lay fundamental groundwork for extending DP to more complex, dynamic, and safety-critical AI domains.

The road ahead demands continued vigilance and innovation. As AI systems become more complex and ubiquitous, the integration of advanced DP techniques will be paramount for mitigating privacy risks, fostering equitable outcomes, and building public trust. These papers collectively paint a picture of a field committed to not just privacy, but effective, practical, and fair privacy for the next generation of AI.

Share this content:

mailbox@3x Differential Privacy: From Convergent Federated Learning to Safeguarding LLMs and Beyond
Hi there 👋

Get a roundup of the latest AI paper digests in a quick, clean weekly email.

Spread the love

Post Comment