Loading…
On Differentially Private Stochastic Convex Optimization with Heavy-tailed Data
In this paper, we consider the problem of designing Differentially Private (DP) algorithms for Stochastic Convex Optimization (SCO) on heavy-tailed data. The irregularity of such data violates some key assumptions used in almost all existing DP-SCO and DP-ERM methods, resulting in failure to provide...
Saved in:
Published in: | arXiv.org 2020-10 |
---|---|
Main Authors: | , , , |
Format: | Article |
Language: | English |
Subjects: | |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
cited_by | |
---|---|
cites | |
container_end_page | |
container_issue | |
container_start_page | |
container_title | arXiv.org |
container_volume | |
creator | Wang, Di Xiao, Hanshen Devadas, Srini Xu, Jinhui |
description | In this paper, we consider the problem of designing Differentially Private (DP) algorithms for Stochastic Convex Optimization (SCO) on heavy-tailed data. The irregularity of such data violates some key assumptions used in almost all existing DP-SCO and DP-ERM methods, resulting in failure to provide the DP guarantees. To better understand this type of challenges, we provide in this paper a comprehensive study of DP-SCO under various settings. First, we consider the case where the loss function is strongly convex and smooth. For this case, we propose a method based on the sample-and-aggregate framework, which has an excess population risk of \(\tilde{O}(\frac{d^3}{n\epsilon^4})\) (after omitting other factors), where \(n\) is the sample size and \(d\) is the dimensionality of the data. Then, we show that with some additional assumptions on the loss functions, it is possible to reduce the \textit{expected} excess population risk to \(\tilde{O}(\frac{ d^2}{ n\epsilon^2 })\). To lift these additional conditions, we also provide a gradient smoothing and trimming based scheme to achieve excess population risks of \(\tilde{O}(\frac{ d^2}{n\epsilon^2})\) and \(\tilde{O}(\frac{d^\frac{2}{3}}{(n\epsilon^2)^\frac{1}{3}})\) for strongly convex and general convex loss functions, respectively, \textit{with high probability}. Experiments suggest that our algorithms can effectively deal with the challenges caused by data irregularity. |
format | article |
fullrecord | <record><control><sourceid>proquest</sourceid><recordid>TN_cdi_proquest_journals_2453523196</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2453523196</sourcerecordid><originalsourceid>FETCH-proquest_journals_24535231963</originalsourceid><addsrcrecordid>eNqNysEKgkAUQNEhCJLyHwZaCzrjWK21cGdQe3nYiE9sxpynZV9fiz6g1V2cu2CekDIK9rEQK-Y714ZhKJKdUEp6rCgMz7Cu9aANIXTdzM8DTkCaX8hWDTjCiqfWTPrFi57wjm8gtIY_kRqea5jmgAA7feMZEGzYsobOaf_XNduejtc0D_rBPkbtqGztOJgvlSJWUgkZHRL53_UBQJg-0A</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2453523196</pqid></control><display><type>article</type><title>On Differentially Private Stochastic Convex Optimization with Heavy-tailed Data</title><source>Publicly Available Content (ProQuest)</source><creator>Wang, Di ; Xiao, Hanshen ; Devadas, Srini ; Xu, Jinhui</creator><creatorcontrib>Wang, Di ; Xiao, Hanshen ; Devadas, Srini ; Xu, Jinhui</creatorcontrib><description>In this paper, we consider the problem of designing Differentially Private (DP) algorithms for Stochastic Convex Optimization (SCO) on heavy-tailed data. The irregularity of such data violates some key assumptions used in almost all existing DP-SCO and DP-ERM methods, resulting in failure to provide the DP guarantees. To better understand this type of challenges, we provide in this paper a comprehensive study of DP-SCO under various settings. First, we consider the case where the loss function is strongly convex and smooth. For this case, we propose a method based on the sample-and-aggregate framework, which has an excess population risk of \(\tilde{O}(\frac{d^3}{n\epsilon^4})\) (after omitting other factors), where \(n\) is the sample size and \(d\) is the dimensionality of the data. Then, we show that with some additional assumptions on the loss functions, it is possible to reduce the \textit{expected} excess population risk to \(\tilde{O}(\frac{ d^2}{ n\epsilon^2 })\). To lift these additional conditions, we also provide a gradient smoothing and trimming based scheme to achieve excess population risks of \(\tilde{O}(\frac{ d^2}{n\epsilon^2})\) and \(\tilde{O}(\frac{d^\frac{2}{3}}{(n\epsilon^2)^\frac{1}{3}})\) for strongly convex and general convex loss functions, respectively, \textit{with high probability}. Experiments suggest that our algorithms can effectively deal with the challenges caused by data irregularity.</description><identifier>EISSN: 2331-8422</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Algorithms ; Computational geometry ; Convex analysis ; Convexity ; Irregularities ; Optimization</subject><ispartof>arXiv.org, 2020-10</ispartof><rights>2020. This work is published under http://arxiv.org/licenses/nonexclusive-distrib/1.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://www.proquest.com/docview/2453523196?pq-origsite=primo$$EHTML$$P50$$Gproquest$$Hfree_for_read</linktohtml><link.rule.ids>780,784,25753,37012,44590</link.rule.ids></links><search><creatorcontrib>Wang, Di</creatorcontrib><creatorcontrib>Xiao, Hanshen</creatorcontrib><creatorcontrib>Devadas, Srini</creatorcontrib><creatorcontrib>Xu, Jinhui</creatorcontrib><title>On Differentially Private Stochastic Convex Optimization with Heavy-tailed Data</title><title>arXiv.org</title><description>In this paper, we consider the problem of designing Differentially Private (DP) algorithms for Stochastic Convex Optimization (SCO) on heavy-tailed data. The irregularity of such data violates some key assumptions used in almost all existing DP-SCO and DP-ERM methods, resulting in failure to provide the DP guarantees. To better understand this type of challenges, we provide in this paper a comprehensive study of DP-SCO under various settings. First, we consider the case where the loss function is strongly convex and smooth. For this case, we propose a method based on the sample-and-aggregate framework, which has an excess population risk of \(\tilde{O}(\frac{d^3}{n\epsilon^4})\) (after omitting other factors), where \(n\) is the sample size and \(d\) is the dimensionality of the data. Then, we show that with some additional assumptions on the loss functions, it is possible to reduce the \textit{expected} excess population risk to \(\tilde{O}(\frac{ d^2}{ n\epsilon^2 })\). To lift these additional conditions, we also provide a gradient smoothing and trimming based scheme to achieve excess population risks of \(\tilde{O}(\frac{ d^2}{n\epsilon^2})\) and \(\tilde{O}(\frac{d^\frac{2}{3}}{(n\epsilon^2)^\frac{1}{3}})\) for strongly convex and general convex loss functions, respectively, \textit{with high probability}. Experiments suggest that our algorithms can effectively deal with the challenges caused by data irregularity.</description><subject>Algorithms</subject><subject>Computational geometry</subject><subject>Convex analysis</subject><subject>Convexity</subject><subject>Irregularities</subject><subject>Optimization</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2020</creationdate><recordtype>article</recordtype><sourceid>PIMPY</sourceid><recordid>eNqNysEKgkAUQNEhCJLyHwZaCzrjWK21cGdQe3nYiE9sxpynZV9fiz6g1V2cu2CekDIK9rEQK-Y714ZhKJKdUEp6rCgMz7Cu9aANIXTdzM8DTkCaX8hWDTjCiqfWTPrFi57wjm8gtIY_kRqea5jmgAA7feMZEGzYsobOaf_XNduejtc0D_rBPkbtqGztOJgvlSJWUgkZHRL53_UBQJg-0A</recordid><startdate>20201021</startdate><enddate>20201021</enddate><creator>Wang, Di</creator><creator>Xiao, Hanshen</creator><creator>Devadas, Srini</creator><creator>Xu, Jinhui</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope></search><sort><creationdate>20201021</creationdate><title>On Differentially Private Stochastic Convex Optimization with Heavy-tailed Data</title><author>Wang, Di ; Xiao, Hanshen ; Devadas, Srini ; Xu, Jinhui</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-proquest_journals_24535231963</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2020</creationdate><topic>Algorithms</topic><topic>Computational geometry</topic><topic>Convex analysis</topic><topic>Convexity</topic><topic>Irregularities</topic><topic>Optimization</topic><toplevel>online_resources</toplevel><creatorcontrib>Wang, Di</creatorcontrib><creatorcontrib>Xiao, Hanshen</creatorcontrib><creatorcontrib>Devadas, Srini</creatorcontrib><creatorcontrib>Xu, Jinhui</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science & Engineering Collection</collection><collection>ProQuest Central (Alumni)</collection><collection>ProQuest Central</collection><collection>ProQuest Central Essentials</collection><collection>AUTh Library subscriptions: ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Publicly Available Content (ProQuest)</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering Collection</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Wang, Di</au><au>Xiao, Hanshen</au><au>Devadas, Srini</au><au>Xu, Jinhui</au><format>book</format><genre>document</genre><ristype>GEN</ristype><atitle>On Differentially Private Stochastic Convex Optimization with Heavy-tailed Data</atitle><jtitle>arXiv.org</jtitle><date>2020-10-21</date><risdate>2020</risdate><eissn>2331-8422</eissn><abstract>In this paper, we consider the problem of designing Differentially Private (DP) algorithms for Stochastic Convex Optimization (SCO) on heavy-tailed data. The irregularity of such data violates some key assumptions used in almost all existing DP-SCO and DP-ERM methods, resulting in failure to provide the DP guarantees. To better understand this type of challenges, we provide in this paper a comprehensive study of DP-SCO under various settings. First, we consider the case where the loss function is strongly convex and smooth. For this case, we propose a method based on the sample-and-aggregate framework, which has an excess population risk of \(\tilde{O}(\frac{d^3}{n\epsilon^4})\) (after omitting other factors), where \(n\) is the sample size and \(d\) is the dimensionality of the data. Then, we show that with some additional assumptions on the loss functions, it is possible to reduce the \textit{expected} excess population risk to \(\tilde{O}(\frac{ d^2}{ n\epsilon^2 })\). To lift these additional conditions, we also provide a gradient smoothing and trimming based scheme to achieve excess population risks of \(\tilde{O}(\frac{ d^2}{n\epsilon^2})\) and \(\tilde{O}(\frac{d^\frac{2}{3}}{(n\epsilon^2)^\frac{1}{3}})\) for strongly convex and general convex loss functions, respectively, \textit{with high probability}. Experiments suggest that our algorithms can effectively deal with the challenges caused by data irregularity.</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | EISSN: 2331-8422 |
ispartof | arXiv.org, 2020-10 |
issn | 2331-8422 |
language | eng |
recordid | cdi_proquest_journals_2453523196 |
source | Publicly Available Content (ProQuest) |
subjects | Algorithms Computational geometry Convex analysis Convexity Irregularities Optimization |
title | On Differentially Private Stochastic Convex Optimization with Heavy-tailed Data |
url | http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-03T23%3A11%3A18IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=document&rft.atitle=On%20Differentially%20Private%20Stochastic%20Convex%20Optimization%20with%20Heavy-tailed%20Data&rft.jtitle=arXiv.org&rft.au=Wang,%20Di&rft.date=2020-10-21&rft.eissn=2331-8422&rft_id=info:doi/&rft_dat=%3Cproquest%3E2453523196%3C/proquest%3E%3Cgrp_id%3Ecdi_FETCH-proquest_journals_24535231963%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2453523196&rft_id=info:pmid/&rfr_iscdi=true |