Fwd: RI Ph.D. Thesis Proposal: Shuli Jiang
Artur Dubrawski
awd at cs.cmu.edu
Tue Nov 12 18:32:13 EST 2024
Of interest to everyone into DAI and coming from a former Autonian.
Artur
---------- Forwarded message ---------
From: Suzanne Muth <lyonsmuth at cmu.edu>
Date: Wed, Nov 13, 2024, 7:23 AM
Subject: RI Ph.D. Thesis Proposal: Shuli Jiang
To: RI People <ri-people at andrew.cmu.edu>
*Date: *22 November 2024
*Time: *9:00 a.m. (ET)
*Location: *NSH 4305
*Zoom Link: *
https://cmu.zoom.us/j/97463407099?pwd=Tp69Yms7hu7kOkgA1jBiQLQPhTyxDe.1
<https://www.google.com/url?q=https://cmu.zoom.us/j/97463407099?pwd%3DTp69Yms7hu7kOkgA1jBiQLQPhTyxDe.1&sa=D&source=calendar&ust=1731745749580401&usg=AOvVaw2uNS6oDx-js-n2WM8Ssyd5>
*Type:* Ph.D. Thesis Proposal
*Who:* Shuli Jiang
*Title:* Communication Efficient and Differentially Private Optimization
*Abstract:*
In recent years, the integration of communication efficiency and
differential privacy in distributed optimization has gained significant
attention, motivated by large-scale applications such as Federated Learning
(FL), where both data privacy and efficient communication are critical.
This thesis explores the development of novel techniques to address these
challenges, with a focus on distributed mean estimation, differentially
private prediction, and private optimization for empirical risk
minimization.
The first part of this work addresses communication-efficient distributed
vector mean estimation, an essential subroutine in distributed optimization
and FL. We propose the Rand-Proj-Spatial family estimator which utilizes
cross-client correlation to reduce the estimation error under fixed
communication cost, by projecting client vectors into a random subspace
using a Subsampled Randomized Hadamard Transform (SRHT). This approach
captures cross-client correlation more effectively, demonstrating
substantial performance gains over conventional sparsification techniques
in various distributed optimization tasks.
The second part of this work focuses on maximizing the privacy-utility
trade-offs in differentially private prediction through majority
ensembling. We introduce the Data-dependent Randomized Response Majority
(DaRRM) framework, which generalizes all private majority ensembling
algorithms through a data-dependent noise function. Based on DaRRM, we
propose a computationally tractable optimization procedure for maximizing
utility under a fixed privacy loss. Empirical results demonstrate DaRRM’s
effectiveness in private label ensembling for image classification, showing
significant utility improvements over existing baselines.
The third part of this work investigates differentially private
optimization in solving empirical risk minimization using shuffled gradient
methods. Unlike conventional private optimizers such as DP-SGD, which
benefits from privacy amplification by subsampling, shuffled gradient
methods face unique challenges in privacy and convergence. We develop a
theoretical framework for analyzing Incremental Gradient (IG) methods, the
most basic form of shuffled gradient methods, that enables noise injection
for privacy and the use of surrogate objectives, introducing a new
dissimilarity metric to measure the difference between true and surrogate
objectives. Leveraging privacy amplification by iteration, we establish the
first empirical excess risk bound for differentially private IG (DP-IG),
and show how interleaving public data in training can further improve
privacy-convergence trade-offs in DP-IG.
Finally, we introduce two proposed works along the line of differentially
private optimization. First, we aim to extend our theoretical framework to
analyze Shuffle Once (SO) and Random Reshuffling (RR), two practical
shuffled gradient methods beyond Incremental Gradient (IG) methods. This
will enable us to understand their private counterparts, DP-SO and DP-RR,
where privacy analysis is more complex due to a lack of understanding on
privacy amplification through shuffling. Second, we plan to extend our
framework from a local to a distributed or decentralized setting to analyze
convergence rates of distributed shuffled gradient methods in both private
and non-private contexts, while also investigating the impact of data
heterogeneity among clients on convergence in this setting.
*Thesis Committee Members:*
Gauri Joshi, Chair
Steven Wu
Zachary Manchester
Swanand Kadhe, IBM Research
A draft of the thesis proposal document can be found here
<https://11hifish.github.io/content/thesis_proposal.pdf>. (available after
Nov.18)
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://mailman.srv.cs.cmu.edu/pipermail/autonlab-users/attachments/20241113/8d0970dd/attachment.html>
More information about the Autonlab-users
mailing list