Loading…

A scalable sparse Cholesky based approach for learning high-dimensional covariance matrices in ordered data

Covariance estimation for high-dimensional datasets is a fundamental problem in machine learning, and has numerous applications. In these high-dimensional settings the number of features or variables p is typically larger than the sample size n . A popular way of tackling this challenge is to induce...

Full description

Saved in:
Bibliographic Details
Published in:Machine learning 2019-12, Vol.108 (12), p.2061-2086
Main Authors: Khare, Kshitij, Oh, Sang-Yun, Rahman, Syed, Rajaratnam, Bala
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Items that cite this one
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by cdi_FETCH-LOGICAL-c363t-b1eb0d8e8ff97aa681b07b2b59163d2768539dc264de32d901785f12c6c694983
cites cdi_FETCH-LOGICAL-c363t-b1eb0d8e8ff97aa681b07b2b59163d2768539dc264de32d901785f12c6c694983
container_end_page 2086
container_issue 12
container_start_page 2061
container_title Machine learning
container_volume 108
creator Khare, Kshitij
Oh, Sang-Yun
Rahman, Syed
Rajaratnam, Bala
description Covariance estimation for high-dimensional datasets is a fundamental problem in machine learning, and has numerous applications. In these high-dimensional settings the number of features or variables p is typically larger than the sample size n . A popular way of tackling this challenge is to induce sparsity in the covariance matrix, its inverse or a relevant transformation. In many applications, the data come with a natural ordering. In such settings, methods inducing sparsity in the Cholesky parameter of the inverse covariance matrix can be quite useful. Such methods are also better positioned to yield a positive definite estimate of the covariance matrix, a critical requirement for several downstream applications. Despite some important advances in this area, a principled approach to general sparse-Cholesky based covariance estimation with both statistical and algorithmic convergence safeguards has been elusive. In particular, the two popular likelihood based methods proposed in the literature either do not lead to a well-defined estimator in high-dimensional settings, or consider only a restrictive class of models. In this paper, we propose a principled and general method for sparse-Cholesky based covariance estimation that aims to overcome some of the shortcomings of current methods, but retains their respective strengths. We obtain a jointly convex formulation for our objective function, and show that it leads to rigorous convergence guarantees and well-defined estimators, even when p > n . Very importantly, the approach always leads to a positive definite and symmetric estimator of the covariance matrix. We establish both high-dimensional estimation and selection consistency, and also demonstrate excellent finite sample performance on simulated/real data.
doi_str_mv 10.1007/s10994-019-05810-5
format article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_2234826101</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2234826101</sourcerecordid><originalsourceid>FETCH-LOGICAL-c363t-b1eb0d8e8ff97aa681b07b2b59163d2768539dc264de32d901785f12c6c694983</originalsourceid><addsrcrecordid>eNp9kEtLxDAUhYMoOI7-AVcB19U8mjRZDoMvENzoOtwm6bQznaYmHWH-vdEK7lxdLpzvcPgQuqbklhJS3SVKtC4LQnVBhKKkECdoQUXF8yvFKVoQpUQhKRPn6CKlLSGESSUXaLfCyUIPde9xGiEmj9dt6H3aHXENyTsM4xgD2BY3IeLeQxy6YYPbbtMWrtv7IXVhgB7b8Amxg8F6vIcpdtYn3A04ROdjbnEwwSU6a6BP_ur3LtH7w_3b-ql4eX18Xq9eCssln4qa-po45VXT6ApAKlqTqma10FRyxyqpBNfOMlk6z5nThFZKNJRZaaUuteJLdDP35uEfB58msw2HmEcmwxgvFZOU0Jxic8rGkFL0jRljt4d4NJSYb6lmlmqyVPMj1YgM8RlKOTxsfPyr_of6Aqkmepc</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2234826101</pqid></control><display><type>article</type><title>A scalable sparse Cholesky based approach for learning high-dimensional covariance matrices in ordered data</title><source>Springer Nature</source><creator>Khare, Kshitij ; Oh, Sang-Yun ; Rahman, Syed ; Rajaratnam, Bala</creator><creatorcontrib>Khare, Kshitij ; Oh, Sang-Yun ; Rahman, Syed ; Rajaratnam, Bala</creatorcontrib><description>Covariance estimation for high-dimensional datasets is a fundamental problem in machine learning, and has numerous applications. In these high-dimensional settings the number of features or variables p is typically larger than the sample size n . A popular way of tackling this challenge is to induce sparsity in the covariance matrix, its inverse or a relevant transformation. In many applications, the data come with a natural ordering. In such settings, methods inducing sparsity in the Cholesky parameter of the inverse covariance matrix can be quite useful. Such methods are also better positioned to yield a positive definite estimate of the covariance matrix, a critical requirement for several downstream applications. Despite some important advances in this area, a principled approach to general sparse-Cholesky based covariance estimation with both statistical and algorithmic convergence safeguards has been elusive. In particular, the two popular likelihood based methods proposed in the literature either do not lead to a well-defined estimator in high-dimensional settings, or consider only a restrictive class of models. In this paper, we propose a principled and general method for sparse-Cholesky based covariance estimation that aims to overcome some of the shortcomings of current methods, but retains their respective strengths. We obtain a jointly convex formulation for our objective function, and show that it leads to rigorous convergence guarantees and well-defined estimators, even when p &gt; n . Very importantly, the approach always leads to a positive definite and symmetric estimator of the covariance matrix. We establish both high-dimensional estimation and selection consistency, and also demonstrate excellent finite sample performance on simulated/real data.</description><identifier>ISSN: 0885-6125</identifier><identifier>EISSN: 1573-0565</identifier><identifier>DOI: 10.1007/s10994-019-05810-5</identifier><language>eng</language><publisher>New York: Springer US</publisher><subject>Artificial Intelligence ; Computer Science ; Computer simulation ; Control ; Convergence ; Covariance matrix ; Economic models ; Machine learning ; Matrix methods ; Mechatronics ; Natural Language Processing (NLP) ; Robotics ; Simulation and Modeling ; Sparsity</subject><ispartof>Machine learning, 2019-12, Vol.108 (12), p.2061-2086</ispartof><rights>The Author(s), under exclusive licence to Springer Science+Business Media LLC, part of Springer Nature 2019</rights><rights>Machine Learning is a copyright of Springer, (2019). All Rights Reserved.</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c363t-b1eb0d8e8ff97aa681b07b2b59163d2768539dc264de32d901785f12c6c694983</citedby><cites>FETCH-LOGICAL-c363t-b1eb0d8e8ff97aa681b07b2b59163d2768539dc264de32d901785f12c6c694983</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,780,784,27923,27924</link.rule.ids></links><search><creatorcontrib>Khare, Kshitij</creatorcontrib><creatorcontrib>Oh, Sang-Yun</creatorcontrib><creatorcontrib>Rahman, Syed</creatorcontrib><creatorcontrib>Rajaratnam, Bala</creatorcontrib><title>A scalable sparse Cholesky based approach for learning high-dimensional covariance matrices in ordered data</title><title>Machine learning</title><addtitle>Mach Learn</addtitle><description>Covariance estimation for high-dimensional datasets is a fundamental problem in machine learning, and has numerous applications. In these high-dimensional settings the number of features or variables p is typically larger than the sample size n . A popular way of tackling this challenge is to induce sparsity in the covariance matrix, its inverse or a relevant transformation. In many applications, the data come with a natural ordering. In such settings, methods inducing sparsity in the Cholesky parameter of the inverse covariance matrix can be quite useful. Such methods are also better positioned to yield a positive definite estimate of the covariance matrix, a critical requirement for several downstream applications. Despite some important advances in this area, a principled approach to general sparse-Cholesky based covariance estimation with both statistical and algorithmic convergence safeguards has been elusive. In particular, the two popular likelihood based methods proposed in the literature either do not lead to a well-defined estimator in high-dimensional settings, or consider only a restrictive class of models. In this paper, we propose a principled and general method for sparse-Cholesky based covariance estimation that aims to overcome some of the shortcomings of current methods, but retains their respective strengths. We obtain a jointly convex formulation for our objective function, and show that it leads to rigorous convergence guarantees and well-defined estimators, even when p &gt; n . Very importantly, the approach always leads to a positive definite and symmetric estimator of the covariance matrix. We establish both high-dimensional estimation and selection consistency, and also demonstrate excellent finite sample performance on simulated/real data.</description><subject>Artificial Intelligence</subject><subject>Computer Science</subject><subject>Computer simulation</subject><subject>Control</subject><subject>Convergence</subject><subject>Covariance matrix</subject><subject>Economic models</subject><subject>Machine learning</subject><subject>Matrix methods</subject><subject>Mechatronics</subject><subject>Natural Language Processing (NLP)</subject><subject>Robotics</subject><subject>Simulation and Modeling</subject><subject>Sparsity</subject><issn>0885-6125</issn><issn>1573-0565</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2019</creationdate><recordtype>article</recordtype><recordid>eNp9kEtLxDAUhYMoOI7-AVcB19U8mjRZDoMvENzoOtwm6bQznaYmHWH-vdEK7lxdLpzvcPgQuqbklhJS3SVKtC4LQnVBhKKkECdoQUXF8yvFKVoQpUQhKRPn6CKlLSGESSUXaLfCyUIPde9xGiEmj9dt6H3aHXENyTsM4xgD2BY3IeLeQxy6YYPbbtMWrtv7IXVhgB7b8Amxg8F6vIcpdtYn3A04ROdjbnEwwSU6a6BP_ur3LtH7w_3b-ql4eX18Xq9eCssln4qa-po45VXT6ApAKlqTqma10FRyxyqpBNfOMlk6z5nThFZKNJRZaaUuteJLdDP35uEfB58msw2HmEcmwxgvFZOU0Jxic8rGkFL0jRljt4d4NJSYb6lmlmqyVPMj1YgM8RlKOTxsfPyr_of6Aqkmepc</recordid><startdate>20191201</startdate><enddate>20191201</enddate><creator>Khare, Kshitij</creator><creator>Oh, Sang-Yun</creator><creator>Rahman, Syed</creator><creator>Rajaratnam, Bala</creator><general>Springer US</general><general>Springer Nature B.V</general><scope>AAYXX</scope><scope>CITATION</scope><scope>3V.</scope><scope>7SC</scope><scope>7XB</scope><scope>88I</scope><scope>8AL</scope><scope>8AO</scope><scope>8FD</scope><scope>8FE</scope><scope>8FG</scope><scope>8FK</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>ARAPS</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>GNUQQ</scope><scope>HCIFZ</scope><scope>JQ2</scope><scope>K7-</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>M0N</scope><scope>M2P</scope><scope>P5Z</scope><scope>P62</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>Q9U</scope></search><sort><creationdate>20191201</creationdate><title>A scalable sparse Cholesky based approach for learning high-dimensional covariance matrices in ordered data</title><author>Khare, Kshitij ; Oh, Sang-Yun ; Rahman, Syed ; Rajaratnam, Bala</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c363t-b1eb0d8e8ff97aa681b07b2b59163d2768539dc264de32d901785f12c6c694983</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2019</creationdate><topic>Artificial Intelligence</topic><topic>Computer Science</topic><topic>Computer simulation</topic><topic>Control</topic><topic>Convergence</topic><topic>Covariance matrix</topic><topic>Economic models</topic><topic>Machine learning</topic><topic>Matrix methods</topic><topic>Mechatronics</topic><topic>Natural Language Processing (NLP)</topic><topic>Robotics</topic><topic>Simulation and Modeling</topic><topic>Sparsity</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Khare, Kshitij</creatorcontrib><creatorcontrib>Oh, Sang-Yun</creatorcontrib><creatorcontrib>Rahman, Syed</creatorcontrib><creatorcontrib>Rajaratnam, Bala</creatorcontrib><collection>CrossRef</collection><collection>ProQuest Central (Corporate)</collection><collection>Computer and Information Systems Abstracts</collection><collection>ProQuest Central (purchase pre-March 2016)</collection><collection>Science Database (Alumni Edition)</collection><collection>Computing Database (Alumni Edition)</collection><collection>ProQuest Pharma Collection</collection><collection>Technology Research Database</collection><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>ProQuest Central (Alumni) (purchase pre-March 2016)</collection><collection>ProQuest Central (Alumni)</collection><collection>ProQuest Central</collection><collection>Advanced Technologies &amp; Aerospace Collection</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>ProQuest Central Student</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Computer Science Collection</collection><collection>Computer Science Database</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>Computing Database</collection><collection>Science Database</collection><collection>Advanced Technologies &amp; Aerospace Database</collection><collection>ProQuest Advanced Technologies &amp; Aerospace Collection</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central Basic</collection><jtitle>Machine learning</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Khare, Kshitij</au><au>Oh, Sang-Yun</au><au>Rahman, Syed</au><au>Rajaratnam, Bala</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>A scalable sparse Cholesky based approach for learning high-dimensional covariance matrices in ordered data</atitle><jtitle>Machine learning</jtitle><stitle>Mach Learn</stitle><date>2019-12-01</date><risdate>2019</risdate><volume>108</volume><issue>12</issue><spage>2061</spage><epage>2086</epage><pages>2061-2086</pages><issn>0885-6125</issn><eissn>1573-0565</eissn><abstract>Covariance estimation for high-dimensional datasets is a fundamental problem in machine learning, and has numerous applications. In these high-dimensional settings the number of features or variables p is typically larger than the sample size n . A popular way of tackling this challenge is to induce sparsity in the covariance matrix, its inverse or a relevant transformation. In many applications, the data come with a natural ordering. In such settings, methods inducing sparsity in the Cholesky parameter of the inverse covariance matrix can be quite useful. Such methods are also better positioned to yield a positive definite estimate of the covariance matrix, a critical requirement for several downstream applications. Despite some important advances in this area, a principled approach to general sparse-Cholesky based covariance estimation with both statistical and algorithmic convergence safeguards has been elusive. In particular, the two popular likelihood based methods proposed in the literature either do not lead to a well-defined estimator in high-dimensional settings, or consider only a restrictive class of models. In this paper, we propose a principled and general method for sparse-Cholesky based covariance estimation that aims to overcome some of the shortcomings of current methods, but retains their respective strengths. We obtain a jointly convex formulation for our objective function, and show that it leads to rigorous convergence guarantees and well-defined estimators, even when p &gt; n . Very importantly, the approach always leads to a positive definite and symmetric estimator of the covariance matrix. We establish both high-dimensional estimation and selection consistency, and also demonstrate excellent finite sample performance on simulated/real data.</abstract><cop>New York</cop><pub>Springer US</pub><doi>10.1007/s10994-019-05810-5</doi><tpages>26</tpages><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 0885-6125
ispartof Machine learning, 2019-12, Vol.108 (12), p.2061-2086
issn 0885-6125
1573-0565
language eng
recordid cdi_proquest_journals_2234826101
source Springer Nature
subjects Artificial Intelligence
Computer Science
Computer simulation
Control
Convergence
Covariance matrix
Economic models
Machine learning
Matrix methods
Mechatronics
Natural Language Processing (NLP)
Robotics
Simulation and Modeling
Sparsity
title A scalable sparse Cholesky based approach for learning high-dimensional covariance matrices in ordered data
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-09T02%3A00%3A08IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=A%20scalable%20sparse%20Cholesky%20based%20approach%20for%20learning%20high-dimensional%20covariance%20matrices%20in%20ordered%20data&rft.jtitle=Machine%20learning&rft.au=Khare,%20Kshitij&rft.date=2019-12-01&rft.volume=108&rft.issue=12&rft.spage=2061&rft.epage=2086&rft.pages=2061-2086&rft.issn=0885-6125&rft.eissn=1573-0565&rft_id=info:doi/10.1007/s10994-019-05810-5&rft_dat=%3Cproquest_cross%3E2234826101%3C/proquest_cross%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c363t-b1eb0d8e8ff97aa681b07b2b59163d2768539dc264de32d901785f12c6c694983%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2234826101&rft_id=info:pmid/&rfr_iscdi=true