Loading…

On Differentially Private Stochastic Convex Optimization with Heavy-tailed Data

In this paper, we consider the problem of designing Differentially Private (DP) algorithms for Stochastic Convex Optimization (SCO) on heavy-tailed data. The irregularity of such data violates some key assumptions used in almost all existing DP-SCO and DP-ERM methods, resulting in failure to provide...

Full description

Saved in:
Bibliographic Details
Published in:arXiv.org 2020-10
Main Authors: Wang, Di, Xiao, Hanshen, Devadas, Srini, Xu, Jinhui
Format: Article
Language:English
Subjects:
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by
cites
container_end_page
container_issue
container_start_page
container_title arXiv.org
container_volume
creator Wang, Di
Xiao, Hanshen
Devadas, Srini
Xu, Jinhui
description In this paper, we consider the problem of designing Differentially Private (DP) algorithms for Stochastic Convex Optimization (SCO) on heavy-tailed data. The irregularity of such data violates some key assumptions used in almost all existing DP-SCO and DP-ERM methods, resulting in failure to provide the DP guarantees. To better understand this type of challenges, we provide in this paper a comprehensive study of DP-SCO under various settings. First, we consider the case where the loss function is strongly convex and smooth. For this case, we propose a method based on the sample-and-aggregate framework, which has an excess population risk of \(\tilde{O}(\frac{d^3}{n\epsilon^4})\) (after omitting other factors), where \(n\) is the sample size and \(d\) is the dimensionality of the data. Then, we show that with some additional assumptions on the loss functions, it is possible to reduce the \textit{expected} excess population risk to \(\tilde{O}(\frac{ d^2}{ n\epsilon^2 })\). To lift these additional conditions, we also provide a gradient smoothing and trimming based scheme to achieve excess population risks of \(\tilde{O}(\frac{ d^2}{n\epsilon^2})\) and \(\tilde{O}(\frac{d^\frac{2}{3}}{(n\epsilon^2)^\frac{1}{3}})\) for strongly convex and general convex loss functions, respectively, \textit{with high probability}. Experiments suggest that our algorithms can effectively deal with the challenges caused by data irregularity.
format article
fullrecord <record><control><sourceid>proquest</sourceid><recordid>TN_cdi_proquest_journals_2453523196</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2453523196</sourcerecordid><originalsourceid>FETCH-proquest_journals_24535231963</originalsourceid><addsrcrecordid>eNqNysEKgkAUQNEhCJLyHwZaCzrjWK21cGdQe3nYiE9sxpynZV9fiz6g1V2cu2CekDIK9rEQK-Y714ZhKJKdUEp6rCgMz7Cu9aANIXTdzM8DTkCaX8hWDTjCiqfWTPrFi57wjm8gtIY_kRqea5jmgAA7feMZEGzYsobOaf_XNduejtc0D_rBPkbtqGztOJgvlSJWUgkZHRL53_UBQJg-0A</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2453523196</pqid></control><display><type>article</type><title>On Differentially Private Stochastic Convex Optimization with Heavy-tailed Data</title><source>Publicly Available Content (ProQuest)</source><creator>Wang, Di ; Xiao, Hanshen ; Devadas, Srini ; Xu, Jinhui</creator><creatorcontrib>Wang, Di ; Xiao, Hanshen ; Devadas, Srini ; Xu, Jinhui</creatorcontrib><description>In this paper, we consider the problem of designing Differentially Private (DP) algorithms for Stochastic Convex Optimization (SCO) on heavy-tailed data. The irregularity of such data violates some key assumptions used in almost all existing DP-SCO and DP-ERM methods, resulting in failure to provide the DP guarantees. To better understand this type of challenges, we provide in this paper a comprehensive study of DP-SCO under various settings. First, we consider the case where the loss function is strongly convex and smooth. For this case, we propose a method based on the sample-and-aggregate framework, which has an excess population risk of \(\tilde{O}(\frac{d^3}{n\epsilon^4})\) (after omitting other factors), where \(n\) is the sample size and \(d\) is the dimensionality of the data. Then, we show that with some additional assumptions on the loss functions, it is possible to reduce the \textit{expected} excess population risk to \(\tilde{O}(\frac{ d^2}{ n\epsilon^2 })\). To lift these additional conditions, we also provide a gradient smoothing and trimming based scheme to achieve excess population risks of \(\tilde{O}(\frac{ d^2}{n\epsilon^2})\) and \(\tilde{O}(\frac{d^\frac{2}{3}}{(n\epsilon^2)^\frac{1}{3}})\) for strongly convex and general convex loss functions, respectively, \textit{with high probability}. Experiments suggest that our algorithms can effectively deal with the challenges caused by data irregularity.</description><identifier>EISSN: 2331-8422</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Algorithms ; Computational geometry ; Convex analysis ; Convexity ; Irregularities ; Optimization</subject><ispartof>arXiv.org, 2020-10</ispartof><rights>2020. This work is published under http://arxiv.org/licenses/nonexclusive-distrib/1.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://www.proquest.com/docview/2453523196?pq-origsite=primo$$EHTML$$P50$$Gproquest$$Hfree_for_read</linktohtml><link.rule.ids>780,784,25753,37012,44590</link.rule.ids></links><search><creatorcontrib>Wang, Di</creatorcontrib><creatorcontrib>Xiao, Hanshen</creatorcontrib><creatorcontrib>Devadas, Srini</creatorcontrib><creatorcontrib>Xu, Jinhui</creatorcontrib><title>On Differentially Private Stochastic Convex Optimization with Heavy-tailed Data</title><title>arXiv.org</title><description>In this paper, we consider the problem of designing Differentially Private (DP) algorithms for Stochastic Convex Optimization (SCO) on heavy-tailed data. The irregularity of such data violates some key assumptions used in almost all existing DP-SCO and DP-ERM methods, resulting in failure to provide the DP guarantees. To better understand this type of challenges, we provide in this paper a comprehensive study of DP-SCO under various settings. First, we consider the case where the loss function is strongly convex and smooth. For this case, we propose a method based on the sample-and-aggregate framework, which has an excess population risk of \(\tilde{O}(\frac{d^3}{n\epsilon^4})\) (after omitting other factors), where \(n\) is the sample size and \(d\) is the dimensionality of the data. Then, we show that with some additional assumptions on the loss functions, it is possible to reduce the \textit{expected} excess population risk to \(\tilde{O}(\frac{ d^2}{ n\epsilon^2 })\). To lift these additional conditions, we also provide a gradient smoothing and trimming based scheme to achieve excess population risks of \(\tilde{O}(\frac{ d^2}{n\epsilon^2})\) and \(\tilde{O}(\frac{d^\frac{2}{3}}{(n\epsilon^2)^\frac{1}{3}})\) for strongly convex and general convex loss functions, respectively, \textit{with high probability}. Experiments suggest that our algorithms can effectively deal with the challenges caused by data irregularity.</description><subject>Algorithms</subject><subject>Computational geometry</subject><subject>Convex analysis</subject><subject>Convexity</subject><subject>Irregularities</subject><subject>Optimization</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2020</creationdate><recordtype>article</recordtype><sourceid>PIMPY</sourceid><recordid>eNqNysEKgkAUQNEhCJLyHwZaCzrjWK21cGdQe3nYiE9sxpynZV9fiz6g1V2cu2CekDIK9rEQK-Y714ZhKJKdUEp6rCgMz7Cu9aANIXTdzM8DTkCaX8hWDTjCiqfWTPrFi57wjm8gtIY_kRqea5jmgAA7feMZEGzYsobOaf_XNduejtc0D_rBPkbtqGztOJgvlSJWUgkZHRL53_UBQJg-0A</recordid><startdate>20201021</startdate><enddate>20201021</enddate><creator>Wang, Di</creator><creator>Xiao, Hanshen</creator><creator>Devadas, Srini</creator><creator>Xu, Jinhui</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope></search><sort><creationdate>20201021</creationdate><title>On Differentially Private Stochastic Convex Optimization with Heavy-tailed Data</title><author>Wang, Di ; Xiao, Hanshen ; Devadas, Srini ; Xu, Jinhui</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-proquest_journals_24535231963</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2020</creationdate><topic>Algorithms</topic><topic>Computational geometry</topic><topic>Convex analysis</topic><topic>Convexity</topic><topic>Irregularities</topic><topic>Optimization</topic><toplevel>online_resources</toplevel><creatorcontrib>Wang, Di</creatorcontrib><creatorcontrib>Xiao, Hanshen</creatorcontrib><creatorcontrib>Devadas, Srini</creatorcontrib><creatorcontrib>Xu, Jinhui</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science &amp; Engineering Collection</collection><collection>ProQuest Central (Alumni)</collection><collection>ProQuest Central</collection><collection>ProQuest Central Essentials</collection><collection>AUTh Library subscriptions: ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Publicly Available Content (ProQuest)</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering Collection</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Wang, Di</au><au>Xiao, Hanshen</au><au>Devadas, Srini</au><au>Xu, Jinhui</au><format>book</format><genre>document</genre><ristype>GEN</ristype><atitle>On Differentially Private Stochastic Convex Optimization with Heavy-tailed Data</atitle><jtitle>arXiv.org</jtitle><date>2020-10-21</date><risdate>2020</risdate><eissn>2331-8422</eissn><abstract>In this paper, we consider the problem of designing Differentially Private (DP) algorithms for Stochastic Convex Optimization (SCO) on heavy-tailed data. The irregularity of such data violates some key assumptions used in almost all existing DP-SCO and DP-ERM methods, resulting in failure to provide the DP guarantees. To better understand this type of challenges, we provide in this paper a comprehensive study of DP-SCO under various settings. First, we consider the case where the loss function is strongly convex and smooth. For this case, we propose a method based on the sample-and-aggregate framework, which has an excess population risk of \(\tilde{O}(\frac{d^3}{n\epsilon^4})\) (after omitting other factors), where \(n\) is the sample size and \(d\) is the dimensionality of the data. Then, we show that with some additional assumptions on the loss functions, it is possible to reduce the \textit{expected} excess population risk to \(\tilde{O}(\frac{ d^2}{ n\epsilon^2 })\). To lift these additional conditions, we also provide a gradient smoothing and trimming based scheme to achieve excess population risks of \(\tilde{O}(\frac{ d^2}{n\epsilon^2})\) and \(\tilde{O}(\frac{d^\frac{2}{3}}{(n\epsilon^2)^\frac{1}{3}})\) for strongly convex and general convex loss functions, respectively, \textit{with high probability}. Experiments suggest that our algorithms can effectively deal with the challenges caused by data irregularity.</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier EISSN: 2331-8422
ispartof arXiv.org, 2020-10
issn 2331-8422
language eng
recordid cdi_proquest_journals_2453523196
source Publicly Available Content (ProQuest)
subjects Algorithms
Computational geometry
Convex analysis
Convexity
Irregularities
Optimization
title On Differentially Private Stochastic Convex Optimization with Heavy-tailed Data
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-03T23%3A11%3A18IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=document&rft.atitle=On%20Differentially%20Private%20Stochastic%20Convex%20Optimization%20with%20Heavy-tailed%20Data&rft.jtitle=arXiv.org&rft.au=Wang,%20Di&rft.date=2020-10-21&rft.eissn=2331-8422&rft_id=info:doi/&rft_dat=%3Cproquest%3E2453523196%3C/proquest%3E%3Cgrp_id%3Ecdi_FETCH-proquest_journals_24535231963%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2453523196&rft_id=info:pmid/&rfr_iscdi=true