Loading…
A scalable sparse Cholesky based approach for learning high-dimensional covariance matrices in ordered data
Covariance estimation for high-dimensional datasets is a fundamental problem in machine learning, and has numerous applications. In these high-dimensional settings the number of features or variables p is typically larger than the sample size n . A popular way of tackling this challenge is to induce...
Saved in:
Published in: | Machine learning 2019-12, Vol.108 (12), p.2061-2086 |
---|---|
Main Authors: | , , , |
Format: | Article |
Language: | English |
Subjects: | |
Citations: | Items that this one cites Items that cite this one |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
cited_by | cdi_FETCH-LOGICAL-c363t-b1eb0d8e8ff97aa681b07b2b59163d2768539dc264de32d901785f12c6c694983 |
---|---|
cites | cdi_FETCH-LOGICAL-c363t-b1eb0d8e8ff97aa681b07b2b59163d2768539dc264de32d901785f12c6c694983 |
container_end_page | 2086 |
container_issue | 12 |
container_start_page | 2061 |
container_title | Machine learning |
container_volume | 108 |
creator | Khare, Kshitij Oh, Sang-Yun Rahman, Syed Rajaratnam, Bala |
description | Covariance estimation for high-dimensional datasets is a fundamental problem in machine learning, and has numerous applications. In these high-dimensional settings the number of features or variables
p
is typically larger than the sample size
n
. A popular way of tackling this challenge is to induce sparsity in the covariance matrix, its inverse or a relevant transformation. In many applications, the data come with a natural ordering. In such settings, methods inducing sparsity in the Cholesky parameter of the inverse covariance matrix can be quite useful. Such methods are also better positioned to yield a positive definite estimate of the covariance matrix, a critical requirement for several downstream applications. Despite some important advances in this area, a principled approach to general sparse-Cholesky based covariance estimation with both statistical and algorithmic convergence safeguards has been elusive. In particular, the two popular likelihood based methods proposed in the literature either do not lead to a well-defined estimator in high-dimensional settings, or consider only a restrictive class of models. In this paper, we propose a principled and general method for sparse-Cholesky based covariance estimation that aims to overcome some of the shortcomings of current methods, but retains their respective strengths. We obtain a
jointly convex
formulation for our objective function, and show that it leads to rigorous convergence guarantees and well-defined estimators, even when
p
>
n
. Very importantly, the approach always leads to a positive definite and symmetric estimator of the covariance matrix. We establish both high-dimensional estimation and selection consistency, and also demonstrate excellent finite sample performance on simulated/real data. |
doi_str_mv | 10.1007/s10994-019-05810-5 |
format | article |
fullrecord | <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_2234826101</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2234826101</sourcerecordid><originalsourceid>FETCH-LOGICAL-c363t-b1eb0d8e8ff97aa681b07b2b59163d2768539dc264de32d901785f12c6c694983</originalsourceid><addsrcrecordid>eNp9kEtLxDAUhYMoOI7-AVcB19U8mjRZDoMvENzoOtwm6bQznaYmHWH-vdEK7lxdLpzvcPgQuqbklhJS3SVKtC4LQnVBhKKkECdoQUXF8yvFKVoQpUQhKRPn6CKlLSGESSUXaLfCyUIPde9xGiEmj9dt6H3aHXENyTsM4xgD2BY3IeLeQxy6YYPbbtMWrtv7IXVhgB7b8Amxg8F6vIcpdtYn3A04ROdjbnEwwSU6a6BP_ur3LtH7w_3b-ql4eX18Xq9eCssln4qa-po45VXT6ApAKlqTqma10FRyxyqpBNfOMlk6z5nThFZKNJRZaaUuteJLdDP35uEfB58msw2HmEcmwxgvFZOU0Jxic8rGkFL0jRljt4d4NJSYb6lmlmqyVPMj1YgM8RlKOTxsfPyr_of6Aqkmepc</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2234826101</pqid></control><display><type>article</type><title>A scalable sparse Cholesky based approach for learning high-dimensional covariance matrices in ordered data</title><source>Springer Nature</source><creator>Khare, Kshitij ; Oh, Sang-Yun ; Rahman, Syed ; Rajaratnam, Bala</creator><creatorcontrib>Khare, Kshitij ; Oh, Sang-Yun ; Rahman, Syed ; Rajaratnam, Bala</creatorcontrib><description>Covariance estimation for high-dimensional datasets is a fundamental problem in machine learning, and has numerous applications. In these high-dimensional settings the number of features or variables
p
is typically larger than the sample size
n
. A popular way of tackling this challenge is to induce sparsity in the covariance matrix, its inverse or a relevant transformation. In many applications, the data come with a natural ordering. In such settings, methods inducing sparsity in the Cholesky parameter of the inverse covariance matrix can be quite useful. Such methods are also better positioned to yield a positive definite estimate of the covariance matrix, a critical requirement for several downstream applications. Despite some important advances in this area, a principled approach to general sparse-Cholesky based covariance estimation with both statistical and algorithmic convergence safeguards has been elusive. In particular, the two popular likelihood based methods proposed in the literature either do not lead to a well-defined estimator in high-dimensional settings, or consider only a restrictive class of models. In this paper, we propose a principled and general method for sparse-Cholesky based covariance estimation that aims to overcome some of the shortcomings of current methods, but retains their respective strengths. We obtain a
jointly convex
formulation for our objective function, and show that it leads to rigorous convergence guarantees and well-defined estimators, even when
p
>
n
. Very importantly, the approach always leads to a positive definite and symmetric estimator of the covariance matrix. We establish both high-dimensional estimation and selection consistency, and also demonstrate excellent finite sample performance on simulated/real data.</description><identifier>ISSN: 0885-6125</identifier><identifier>EISSN: 1573-0565</identifier><identifier>DOI: 10.1007/s10994-019-05810-5</identifier><language>eng</language><publisher>New York: Springer US</publisher><subject>Artificial Intelligence ; Computer Science ; Computer simulation ; Control ; Convergence ; Covariance matrix ; Economic models ; Machine learning ; Matrix methods ; Mechatronics ; Natural Language Processing (NLP) ; Robotics ; Simulation and Modeling ; Sparsity</subject><ispartof>Machine learning, 2019-12, Vol.108 (12), p.2061-2086</ispartof><rights>The Author(s), under exclusive licence to Springer Science+Business Media LLC, part of Springer Nature 2019</rights><rights>Machine Learning is a copyright of Springer, (2019). All Rights Reserved.</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c363t-b1eb0d8e8ff97aa681b07b2b59163d2768539dc264de32d901785f12c6c694983</citedby><cites>FETCH-LOGICAL-c363t-b1eb0d8e8ff97aa681b07b2b59163d2768539dc264de32d901785f12c6c694983</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,780,784,27923,27924</link.rule.ids></links><search><creatorcontrib>Khare, Kshitij</creatorcontrib><creatorcontrib>Oh, Sang-Yun</creatorcontrib><creatorcontrib>Rahman, Syed</creatorcontrib><creatorcontrib>Rajaratnam, Bala</creatorcontrib><title>A scalable sparse Cholesky based approach for learning high-dimensional covariance matrices in ordered data</title><title>Machine learning</title><addtitle>Mach Learn</addtitle><description>Covariance estimation for high-dimensional datasets is a fundamental problem in machine learning, and has numerous applications. In these high-dimensional settings the number of features or variables
p
is typically larger than the sample size
n
. A popular way of tackling this challenge is to induce sparsity in the covariance matrix, its inverse or a relevant transformation. In many applications, the data come with a natural ordering. In such settings, methods inducing sparsity in the Cholesky parameter of the inverse covariance matrix can be quite useful. Such methods are also better positioned to yield a positive definite estimate of the covariance matrix, a critical requirement for several downstream applications. Despite some important advances in this area, a principled approach to general sparse-Cholesky based covariance estimation with both statistical and algorithmic convergence safeguards has been elusive. In particular, the two popular likelihood based methods proposed in the literature either do not lead to a well-defined estimator in high-dimensional settings, or consider only a restrictive class of models. In this paper, we propose a principled and general method for sparse-Cholesky based covariance estimation that aims to overcome some of the shortcomings of current methods, but retains their respective strengths. We obtain a
jointly convex
formulation for our objective function, and show that it leads to rigorous convergence guarantees and well-defined estimators, even when
p
>
n
. Very importantly, the approach always leads to a positive definite and symmetric estimator of the covariance matrix. We establish both high-dimensional estimation and selection consistency, and also demonstrate excellent finite sample performance on simulated/real data.</description><subject>Artificial Intelligence</subject><subject>Computer Science</subject><subject>Computer simulation</subject><subject>Control</subject><subject>Convergence</subject><subject>Covariance matrix</subject><subject>Economic models</subject><subject>Machine learning</subject><subject>Matrix methods</subject><subject>Mechatronics</subject><subject>Natural Language Processing (NLP)</subject><subject>Robotics</subject><subject>Simulation and Modeling</subject><subject>Sparsity</subject><issn>0885-6125</issn><issn>1573-0565</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2019</creationdate><recordtype>article</recordtype><recordid>eNp9kEtLxDAUhYMoOI7-AVcB19U8mjRZDoMvENzoOtwm6bQznaYmHWH-vdEK7lxdLpzvcPgQuqbklhJS3SVKtC4LQnVBhKKkECdoQUXF8yvFKVoQpUQhKRPn6CKlLSGESSUXaLfCyUIPde9xGiEmj9dt6H3aHXENyTsM4xgD2BY3IeLeQxy6YYPbbtMWrtv7IXVhgB7b8Amxg8F6vIcpdtYn3A04ROdjbnEwwSU6a6BP_ur3LtH7w_3b-ql4eX18Xq9eCssln4qa-po45VXT6ApAKlqTqma10FRyxyqpBNfOMlk6z5nThFZKNJRZaaUuteJLdDP35uEfB58msw2HmEcmwxgvFZOU0Jxic8rGkFL0jRljt4d4NJSYb6lmlmqyVPMj1YgM8RlKOTxsfPyr_of6Aqkmepc</recordid><startdate>20191201</startdate><enddate>20191201</enddate><creator>Khare, Kshitij</creator><creator>Oh, Sang-Yun</creator><creator>Rahman, Syed</creator><creator>Rajaratnam, Bala</creator><general>Springer US</general><general>Springer Nature B.V</general><scope>AAYXX</scope><scope>CITATION</scope><scope>3V.</scope><scope>7SC</scope><scope>7XB</scope><scope>88I</scope><scope>8AL</scope><scope>8AO</scope><scope>8FD</scope><scope>8FE</scope><scope>8FG</scope><scope>8FK</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>ARAPS</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>GNUQQ</scope><scope>HCIFZ</scope><scope>JQ2</scope><scope>K7-</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>M0N</scope><scope>M2P</scope><scope>P5Z</scope><scope>P62</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>Q9U</scope></search><sort><creationdate>20191201</creationdate><title>A scalable sparse Cholesky based approach for learning high-dimensional covariance matrices in ordered data</title><author>Khare, Kshitij ; Oh, Sang-Yun ; Rahman, Syed ; Rajaratnam, Bala</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c363t-b1eb0d8e8ff97aa681b07b2b59163d2768539dc264de32d901785f12c6c694983</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2019</creationdate><topic>Artificial Intelligence</topic><topic>Computer Science</topic><topic>Computer simulation</topic><topic>Control</topic><topic>Convergence</topic><topic>Covariance matrix</topic><topic>Economic models</topic><topic>Machine learning</topic><topic>Matrix methods</topic><topic>Mechatronics</topic><topic>Natural Language Processing (NLP)</topic><topic>Robotics</topic><topic>Simulation and Modeling</topic><topic>Sparsity</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Khare, Kshitij</creatorcontrib><creatorcontrib>Oh, Sang-Yun</creatorcontrib><creatorcontrib>Rahman, Syed</creatorcontrib><creatorcontrib>Rajaratnam, Bala</creatorcontrib><collection>CrossRef</collection><collection>ProQuest Central (Corporate)</collection><collection>Computer and Information Systems Abstracts</collection><collection>ProQuest Central (purchase pre-March 2016)</collection><collection>Science Database (Alumni Edition)</collection><collection>Computing Database (Alumni Edition)</collection><collection>ProQuest Pharma Collection</collection><collection>Technology Research Database</collection><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>ProQuest Central (Alumni) (purchase pre-March 2016)</collection><collection>ProQuest Central (Alumni)</collection><collection>ProQuest Central</collection><collection>Advanced Technologies & Aerospace Collection</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>ProQuest Central Student</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Computer Science Collection</collection><collection>Computer Science Database</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>Computing Database</collection><collection>Science Database</collection><collection>Advanced Technologies & Aerospace Database</collection><collection>ProQuest Advanced Technologies & Aerospace Collection</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central Basic</collection><jtitle>Machine learning</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Khare, Kshitij</au><au>Oh, Sang-Yun</au><au>Rahman, Syed</au><au>Rajaratnam, Bala</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>A scalable sparse Cholesky based approach for learning high-dimensional covariance matrices in ordered data</atitle><jtitle>Machine learning</jtitle><stitle>Mach Learn</stitle><date>2019-12-01</date><risdate>2019</risdate><volume>108</volume><issue>12</issue><spage>2061</spage><epage>2086</epage><pages>2061-2086</pages><issn>0885-6125</issn><eissn>1573-0565</eissn><abstract>Covariance estimation for high-dimensional datasets is a fundamental problem in machine learning, and has numerous applications. In these high-dimensional settings the number of features or variables
p
is typically larger than the sample size
n
. A popular way of tackling this challenge is to induce sparsity in the covariance matrix, its inverse or a relevant transformation. In many applications, the data come with a natural ordering. In such settings, methods inducing sparsity in the Cholesky parameter of the inverse covariance matrix can be quite useful. Such methods are also better positioned to yield a positive definite estimate of the covariance matrix, a critical requirement for several downstream applications. Despite some important advances in this area, a principled approach to general sparse-Cholesky based covariance estimation with both statistical and algorithmic convergence safeguards has been elusive. In particular, the two popular likelihood based methods proposed in the literature either do not lead to a well-defined estimator in high-dimensional settings, or consider only a restrictive class of models. In this paper, we propose a principled and general method for sparse-Cholesky based covariance estimation that aims to overcome some of the shortcomings of current methods, but retains their respective strengths. We obtain a
jointly convex
formulation for our objective function, and show that it leads to rigorous convergence guarantees and well-defined estimators, even when
p
>
n
. Very importantly, the approach always leads to a positive definite and symmetric estimator of the covariance matrix. We establish both high-dimensional estimation and selection consistency, and also demonstrate excellent finite sample performance on simulated/real data.</abstract><cop>New York</cop><pub>Springer US</pub><doi>10.1007/s10994-019-05810-5</doi><tpages>26</tpages><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | ISSN: 0885-6125 |
ispartof | Machine learning, 2019-12, Vol.108 (12), p.2061-2086 |
issn | 0885-6125 1573-0565 |
language | eng |
recordid | cdi_proquest_journals_2234826101 |
source | Springer Nature |
subjects | Artificial Intelligence Computer Science Computer simulation Control Convergence Covariance matrix Economic models Machine learning Matrix methods Mechatronics Natural Language Processing (NLP) Robotics Simulation and Modeling Sparsity |
title | A scalable sparse Cholesky based approach for learning high-dimensional covariance matrices in ordered data |
url | http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-09T02%3A00%3A08IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=A%20scalable%20sparse%20Cholesky%20based%20approach%20for%20learning%20high-dimensional%20covariance%20matrices%20in%20ordered%20data&rft.jtitle=Machine%20learning&rft.au=Khare,%20Kshitij&rft.date=2019-12-01&rft.volume=108&rft.issue=12&rft.spage=2061&rft.epage=2086&rft.pages=2061-2086&rft.issn=0885-6125&rft.eissn=1573-0565&rft_id=info:doi/10.1007/s10994-019-05810-5&rft_dat=%3Cproquest_cross%3E2234826101%3C/proquest_cross%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c363t-b1eb0d8e8ff97aa681b07b2b59163d2768539dc264de32d901785f12c6c694983%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2234826101&rft_id=info:pmid/&rfr_iscdi=true |