Loading…
Bayesian Active Learning for Discrete Latent Variable Models
Active learning seeks to reduce the amount of data required to fit the parameters of a model, thus forming an important class of techniques in modern machine learning. However, past work on active learning has largely overlooked latent variable models, which play a vital role in neuroscience, psycho...
Saved in:
Published in: | arXiv.org 2023-06 |
---|---|
Main Authors: | , , |
Format: | Article |
Language: | English |
Subjects: | |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
cited_by | |
---|---|
cites | |
container_end_page | |
container_issue | |
container_start_page | |
container_title | arXiv.org |
container_volume | |
creator | Jha, Aditi Ashwood, Zoe C Pillow, Jonathan W |
description | Active learning seeks to reduce the amount of data required to fit the parameters of a model, thus forming an important class of techniques in modern machine learning. However, past work on active learning has largely overlooked latent variable models, which play a vital role in neuroscience, psychology, and a variety of other engineering and scientific disciplines. Here we address this gap by proposing a novel framework for maximum-mutual-information input selection for discrete latent variable regression models. We first apply our method to a class of models known as "mixtures of linear regressions" (MLR). While it is well known that active learning confers no advantage for linear-Gaussian regression models, we use Fisher information to show analytically that active learning can nevertheless achieve large gains for mixtures of such models, and we validate this improvement using both simulations and real-world data. We then consider a powerful class of temporally structured latent variable models given by a Hidden Markov Model (HMM) with generalized linear model (GLM) observations, which has recently been used to identify discrete states from animal decision-making data. We show that our method substantially reduces the amount of data needed to fit GLM-HMM, and outperforms a variety of approximate methods based on variational and amortized inference. Infomax learning for latent variable models thus offers a powerful for characterizing temporally structured latent states, with a wide variety of applications in neuroscience and beyond. |
doi_str_mv | 10.48550/arxiv.2202.13426 |
format | article |
fullrecord | <record><control><sourceid>proquest</sourceid><recordid>TN_cdi_proquest_journals_2634666157</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2634666157</sourcerecordid><originalsourceid>FETCH-LOGICAL-a956-bd7c284281d5dac14e8488554daaf48a51002653720ccf10278ef61d99d558683</originalsourceid><addsrcrecordid>eNotjs1Kw0AURgdBsNQ-gLsB14kzd34yATe1ahUi3ZRuy-3MjUwJic6kRd_eiK4-OIvvHMZupCi1M0bcYfqK5xJAQCmVBnvBZqCULJwGuGKLnI9CCLAVGKNm7P4BvylH7PnSj_FMvCFMfezfeTsk_hizTzROFEfqR77DFPHQEX8bAnX5ml222GVa_O-cbZ-ftquXotmsX1fLpsDa2OIQKg-T3clgAnqpyWk3peqA2GqHRv72GFWB8L6VAipHrZWhroMxzjo1Z7d_tx9p-DxRHvfH4ZT6ybgHq7S1VppK_QB-QUfo</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2634666157</pqid></control><display><type>article</type><title>Bayesian Active Learning for Discrete Latent Variable Models</title><source>Publicly Available Content Database (Proquest) (PQ_SDU_P3)</source><creator>Jha, Aditi ; Ashwood, Zoe C ; Pillow, Jonathan W</creator><creatorcontrib>Jha, Aditi ; Ashwood, Zoe C ; Pillow, Jonathan W</creatorcontrib><description>Active learning seeks to reduce the amount of data required to fit the parameters of a model, thus forming an important class of techniques in modern machine learning. However, past work on active learning has largely overlooked latent variable models, which play a vital role in neuroscience, psychology, and a variety of other engineering and scientific disciplines. Here we address this gap by proposing a novel framework for maximum-mutual-information input selection for discrete latent variable regression models. We first apply our method to a class of models known as "mixtures of linear regressions" (MLR). While it is well known that active learning confers no advantage for linear-Gaussian regression models, we use Fisher information to show analytically that active learning can nevertheless achieve large gains for mixtures of such models, and we validate this improvement using both simulations and real-world data. We then consider a powerful class of temporally structured latent variable models given by a Hidden Markov Model (HMM) with generalized linear model (GLM) observations, which has recently been used to identify discrete states from animal decision-making data. We show that our method substantially reduces the amount of data needed to fit GLM-HMM, and outperforms a variety of approximate methods based on variational and amortized inference. Infomax learning for latent variable models thus offers a powerful for characterizing temporally structured latent states, with a wide variety of applications in neuroscience and beyond.</description><identifier>EISSN: 2331-8422</identifier><identifier>DOI: 10.48550/arxiv.2202.13426</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Active learning ; Approximation ; Least squares method ; Machine learning ; Markov chains ; Neurosciences ; Psychology ; Regression models</subject><ispartof>arXiv.org, 2023-06</ispartof><rights>2023. This work is published under http://arxiv.org/licenses/nonexclusive-distrib/1.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://www.proquest.com/docview/2634666157?pq-origsite=primo$$EHTML$$P50$$Gproquest$$Hfree_for_read</linktohtml><link.rule.ids>780,784,25751,27923,37010,44588</link.rule.ids></links><search><creatorcontrib>Jha, Aditi</creatorcontrib><creatorcontrib>Ashwood, Zoe C</creatorcontrib><creatorcontrib>Pillow, Jonathan W</creatorcontrib><title>Bayesian Active Learning for Discrete Latent Variable Models</title><title>arXiv.org</title><description>Active learning seeks to reduce the amount of data required to fit the parameters of a model, thus forming an important class of techniques in modern machine learning. However, past work on active learning has largely overlooked latent variable models, which play a vital role in neuroscience, psychology, and a variety of other engineering and scientific disciplines. Here we address this gap by proposing a novel framework for maximum-mutual-information input selection for discrete latent variable regression models. We first apply our method to a class of models known as "mixtures of linear regressions" (MLR). While it is well known that active learning confers no advantage for linear-Gaussian regression models, we use Fisher information to show analytically that active learning can nevertheless achieve large gains for mixtures of such models, and we validate this improvement using both simulations and real-world data. We then consider a powerful class of temporally structured latent variable models given by a Hidden Markov Model (HMM) with generalized linear model (GLM) observations, which has recently been used to identify discrete states from animal decision-making data. We show that our method substantially reduces the amount of data needed to fit GLM-HMM, and outperforms a variety of approximate methods based on variational and amortized inference. Infomax learning for latent variable models thus offers a powerful for characterizing temporally structured latent states, with a wide variety of applications in neuroscience and beyond.</description><subject>Active learning</subject><subject>Approximation</subject><subject>Least squares method</subject><subject>Machine learning</subject><subject>Markov chains</subject><subject>Neurosciences</subject><subject>Psychology</subject><subject>Regression models</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>PIMPY</sourceid><recordid>eNotjs1Kw0AURgdBsNQ-gLsB14kzd34yATe1ahUi3ZRuy-3MjUwJic6kRd_eiK4-OIvvHMZupCi1M0bcYfqK5xJAQCmVBnvBZqCULJwGuGKLnI9CCLAVGKNm7P4BvylH7PnSj_FMvCFMfezfeTsk_hizTzROFEfqR77DFPHQEX8bAnX5ml222GVa_O-cbZ-ftquXotmsX1fLpsDa2OIQKg-T3clgAnqpyWk3peqA2GqHRv72GFWB8L6VAipHrZWhroMxzjo1Z7d_tx9p-DxRHvfH4ZT6ybgHq7S1VppK_QB-QUfo</recordid><startdate>20230602</startdate><enddate>20230602</enddate><creator>Jha, Aditi</creator><creator>Ashwood, Zoe C</creator><creator>Pillow, Jonathan W</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope></search><sort><creationdate>20230602</creationdate><title>Bayesian Active Learning for Discrete Latent Variable Models</title><author>Jha, Aditi ; Ashwood, Zoe C ; Pillow, Jonathan W</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a956-bd7c284281d5dac14e8488554daaf48a51002653720ccf10278ef61d99d558683</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Active learning</topic><topic>Approximation</topic><topic>Least squares method</topic><topic>Machine learning</topic><topic>Markov chains</topic><topic>Neurosciences</topic><topic>Psychology</topic><topic>Regression models</topic><toplevel>online_resources</toplevel><creatorcontrib>Jha, Aditi</creatorcontrib><creatorcontrib>Ashwood, Zoe C</creatorcontrib><creatorcontrib>Pillow, Jonathan W</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science & Engineering Collection</collection><collection>ProQuest Central (Alumni)</collection><collection>ProQuest Central</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Publicly Available Content Database (Proquest) (PQ_SDU_P3)</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering Collection</collection><jtitle>arXiv.org</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Jha, Aditi</au><au>Ashwood, Zoe C</au><au>Pillow, Jonathan W</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Bayesian Active Learning for Discrete Latent Variable Models</atitle><jtitle>arXiv.org</jtitle><date>2023-06-02</date><risdate>2023</risdate><eissn>2331-8422</eissn><abstract>Active learning seeks to reduce the amount of data required to fit the parameters of a model, thus forming an important class of techniques in modern machine learning. However, past work on active learning has largely overlooked latent variable models, which play a vital role in neuroscience, psychology, and a variety of other engineering and scientific disciplines. Here we address this gap by proposing a novel framework for maximum-mutual-information input selection for discrete latent variable regression models. We first apply our method to a class of models known as "mixtures of linear regressions" (MLR). While it is well known that active learning confers no advantage for linear-Gaussian regression models, we use Fisher information to show analytically that active learning can nevertheless achieve large gains for mixtures of such models, and we validate this improvement using both simulations and real-world data. We then consider a powerful class of temporally structured latent variable models given by a Hidden Markov Model (HMM) with generalized linear model (GLM) observations, which has recently been used to identify discrete states from animal decision-making data. We show that our method substantially reduces the amount of data needed to fit GLM-HMM, and outperforms a variety of approximate methods based on variational and amortized inference. Infomax learning for latent variable models thus offers a powerful for characterizing temporally structured latent states, with a wide variety of applications in neuroscience and beyond.</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><doi>10.48550/arxiv.2202.13426</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | EISSN: 2331-8422 |
ispartof | arXiv.org, 2023-06 |
issn | 2331-8422 |
language | eng |
recordid | cdi_proquest_journals_2634666157 |
source | Publicly Available Content Database (Proquest) (PQ_SDU_P3) |
subjects | Active learning Approximation Least squares method Machine learning Markov chains Neurosciences Psychology Regression models |
title | Bayesian Active Learning for Discrete Latent Variable Models |
url | http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-14T13%3A18%3A29IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Bayesian%20Active%20Learning%20for%20Discrete%20Latent%20Variable%20Models&rft.jtitle=arXiv.org&rft.au=Jha,%20Aditi&rft.date=2023-06-02&rft.eissn=2331-8422&rft_id=info:doi/10.48550/arxiv.2202.13426&rft_dat=%3Cproquest%3E2634666157%3C/proquest%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-a956-bd7c284281d5dac14e8488554daaf48a51002653720ccf10278ef61d99d558683%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2634666157&rft_id=info:pmid/&rfr_iscdi=true |