Loading…

Bayesian Active Learning for Discrete Latent Variable Models

Active learning seeks to reduce the amount of data required to fit the parameters of a model, thus forming an important class of techniques in modern machine learning. However, past work on active learning has largely overlooked latent variable models, which play a vital role in neuroscience, psycho...

Full description

Saved in:
Bibliographic Details
Published in:arXiv.org 2023-06
Main Authors: Jha, Aditi, Ashwood, Zoe C, Pillow, Jonathan W
Format: Article
Language:English
Subjects:
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by
cites
container_end_page
container_issue
container_start_page
container_title arXiv.org
container_volume
creator Jha, Aditi
Ashwood, Zoe C
Pillow, Jonathan W
description Active learning seeks to reduce the amount of data required to fit the parameters of a model, thus forming an important class of techniques in modern machine learning. However, past work on active learning has largely overlooked latent variable models, which play a vital role in neuroscience, psychology, and a variety of other engineering and scientific disciplines. Here we address this gap by proposing a novel framework for maximum-mutual-information input selection for discrete latent variable regression models. We first apply our method to a class of models known as "mixtures of linear regressions" (MLR). While it is well known that active learning confers no advantage for linear-Gaussian regression models, we use Fisher information to show analytically that active learning can nevertheless achieve large gains for mixtures of such models, and we validate this improvement using both simulations and real-world data. We then consider a powerful class of temporally structured latent variable models given by a Hidden Markov Model (HMM) with generalized linear model (GLM) observations, which has recently been used to identify discrete states from animal decision-making data. We show that our method substantially reduces the amount of data needed to fit GLM-HMM, and outperforms a variety of approximate methods based on variational and amortized inference. Infomax learning for latent variable models thus offers a powerful for characterizing temporally structured latent states, with a wide variety of applications in neuroscience and beyond.
doi_str_mv 10.48550/arxiv.2202.13426
format article
fullrecord <record><control><sourceid>proquest</sourceid><recordid>TN_cdi_proquest_journals_2634666157</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2634666157</sourcerecordid><originalsourceid>FETCH-LOGICAL-a956-bd7c284281d5dac14e8488554daaf48a51002653720ccf10278ef61d99d558683</originalsourceid><addsrcrecordid>eNotjs1Kw0AURgdBsNQ-gLsB14kzd34yATe1ahUi3ZRuy-3MjUwJic6kRd_eiK4-OIvvHMZupCi1M0bcYfqK5xJAQCmVBnvBZqCULJwGuGKLnI9CCLAVGKNm7P4BvylH7PnSj_FMvCFMfezfeTsk_hizTzROFEfqR77DFPHQEX8bAnX5ml222GVa_O-cbZ-ftquXotmsX1fLpsDa2OIQKg-T3clgAnqpyWk3peqA2GqHRv72GFWB8L6VAipHrZWhroMxzjo1Z7d_tx9p-DxRHvfH4ZT6ybgHq7S1VppK_QB-QUfo</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2634666157</pqid></control><display><type>article</type><title>Bayesian Active Learning for Discrete Latent Variable Models</title><source>Publicly Available Content Database (Proquest) (PQ_SDU_P3)</source><creator>Jha, Aditi ; Ashwood, Zoe C ; Pillow, Jonathan W</creator><creatorcontrib>Jha, Aditi ; Ashwood, Zoe C ; Pillow, Jonathan W</creatorcontrib><description>Active learning seeks to reduce the amount of data required to fit the parameters of a model, thus forming an important class of techniques in modern machine learning. However, past work on active learning has largely overlooked latent variable models, which play a vital role in neuroscience, psychology, and a variety of other engineering and scientific disciplines. Here we address this gap by proposing a novel framework for maximum-mutual-information input selection for discrete latent variable regression models. We first apply our method to a class of models known as "mixtures of linear regressions" (MLR). While it is well known that active learning confers no advantage for linear-Gaussian regression models, we use Fisher information to show analytically that active learning can nevertheless achieve large gains for mixtures of such models, and we validate this improvement using both simulations and real-world data. We then consider a powerful class of temporally structured latent variable models given by a Hidden Markov Model (HMM) with generalized linear model (GLM) observations, which has recently been used to identify discrete states from animal decision-making data. We show that our method substantially reduces the amount of data needed to fit GLM-HMM, and outperforms a variety of approximate methods based on variational and amortized inference. Infomax learning for latent variable models thus offers a powerful for characterizing temporally structured latent states, with a wide variety of applications in neuroscience and beyond.</description><identifier>EISSN: 2331-8422</identifier><identifier>DOI: 10.48550/arxiv.2202.13426</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Active learning ; Approximation ; Least squares method ; Machine learning ; Markov chains ; Neurosciences ; Psychology ; Regression models</subject><ispartof>arXiv.org, 2023-06</ispartof><rights>2023. This work is published under http://arxiv.org/licenses/nonexclusive-distrib/1.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://www.proquest.com/docview/2634666157?pq-origsite=primo$$EHTML$$P50$$Gproquest$$Hfree_for_read</linktohtml><link.rule.ids>780,784,25751,27923,37010,44588</link.rule.ids></links><search><creatorcontrib>Jha, Aditi</creatorcontrib><creatorcontrib>Ashwood, Zoe C</creatorcontrib><creatorcontrib>Pillow, Jonathan W</creatorcontrib><title>Bayesian Active Learning for Discrete Latent Variable Models</title><title>arXiv.org</title><description>Active learning seeks to reduce the amount of data required to fit the parameters of a model, thus forming an important class of techniques in modern machine learning. However, past work on active learning has largely overlooked latent variable models, which play a vital role in neuroscience, psychology, and a variety of other engineering and scientific disciplines. Here we address this gap by proposing a novel framework for maximum-mutual-information input selection for discrete latent variable regression models. We first apply our method to a class of models known as "mixtures of linear regressions" (MLR). While it is well known that active learning confers no advantage for linear-Gaussian regression models, we use Fisher information to show analytically that active learning can nevertheless achieve large gains for mixtures of such models, and we validate this improvement using both simulations and real-world data. We then consider a powerful class of temporally structured latent variable models given by a Hidden Markov Model (HMM) with generalized linear model (GLM) observations, which has recently been used to identify discrete states from animal decision-making data. We show that our method substantially reduces the amount of data needed to fit GLM-HMM, and outperforms a variety of approximate methods based on variational and amortized inference. Infomax learning for latent variable models thus offers a powerful for characterizing temporally structured latent states, with a wide variety of applications in neuroscience and beyond.</description><subject>Active learning</subject><subject>Approximation</subject><subject>Least squares method</subject><subject>Machine learning</subject><subject>Markov chains</subject><subject>Neurosciences</subject><subject>Psychology</subject><subject>Regression models</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>PIMPY</sourceid><recordid>eNotjs1Kw0AURgdBsNQ-gLsB14kzd34yATe1ahUi3ZRuy-3MjUwJic6kRd_eiK4-OIvvHMZupCi1M0bcYfqK5xJAQCmVBnvBZqCULJwGuGKLnI9CCLAVGKNm7P4BvylH7PnSj_FMvCFMfezfeTsk_hizTzROFEfqR77DFPHQEX8bAnX5ml222GVa_O-cbZ-ftquXotmsX1fLpsDa2OIQKg-T3clgAnqpyWk3peqA2GqHRv72GFWB8L6VAipHrZWhroMxzjo1Z7d_tx9p-DxRHvfH4ZT6ybgHq7S1VppK_QB-QUfo</recordid><startdate>20230602</startdate><enddate>20230602</enddate><creator>Jha, Aditi</creator><creator>Ashwood, Zoe C</creator><creator>Pillow, Jonathan W</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope></search><sort><creationdate>20230602</creationdate><title>Bayesian Active Learning for Discrete Latent Variable Models</title><author>Jha, Aditi ; Ashwood, Zoe C ; Pillow, Jonathan W</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a956-bd7c284281d5dac14e8488554daaf48a51002653720ccf10278ef61d99d558683</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Active learning</topic><topic>Approximation</topic><topic>Least squares method</topic><topic>Machine learning</topic><topic>Markov chains</topic><topic>Neurosciences</topic><topic>Psychology</topic><topic>Regression models</topic><toplevel>online_resources</toplevel><creatorcontrib>Jha, Aditi</creatorcontrib><creatorcontrib>Ashwood, Zoe C</creatorcontrib><creatorcontrib>Pillow, Jonathan W</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science &amp; Engineering Collection</collection><collection>ProQuest Central (Alumni)</collection><collection>ProQuest Central</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Publicly Available Content Database (Proquest) (PQ_SDU_P3)</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering Collection</collection><jtitle>arXiv.org</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Jha, Aditi</au><au>Ashwood, Zoe C</au><au>Pillow, Jonathan W</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Bayesian Active Learning for Discrete Latent Variable Models</atitle><jtitle>arXiv.org</jtitle><date>2023-06-02</date><risdate>2023</risdate><eissn>2331-8422</eissn><abstract>Active learning seeks to reduce the amount of data required to fit the parameters of a model, thus forming an important class of techniques in modern machine learning. However, past work on active learning has largely overlooked latent variable models, which play a vital role in neuroscience, psychology, and a variety of other engineering and scientific disciplines. Here we address this gap by proposing a novel framework for maximum-mutual-information input selection for discrete latent variable regression models. We first apply our method to a class of models known as "mixtures of linear regressions" (MLR). While it is well known that active learning confers no advantage for linear-Gaussian regression models, we use Fisher information to show analytically that active learning can nevertheless achieve large gains for mixtures of such models, and we validate this improvement using both simulations and real-world data. We then consider a powerful class of temporally structured latent variable models given by a Hidden Markov Model (HMM) with generalized linear model (GLM) observations, which has recently been used to identify discrete states from animal decision-making data. We show that our method substantially reduces the amount of data needed to fit GLM-HMM, and outperforms a variety of approximate methods based on variational and amortized inference. Infomax learning for latent variable models thus offers a powerful for characterizing temporally structured latent states, with a wide variety of applications in neuroscience and beyond.</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><doi>10.48550/arxiv.2202.13426</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier EISSN: 2331-8422
ispartof arXiv.org, 2023-06
issn 2331-8422
language eng
recordid cdi_proquest_journals_2634666157
source Publicly Available Content Database (Proquest) (PQ_SDU_P3)
subjects Active learning
Approximation
Least squares method
Machine learning
Markov chains
Neurosciences
Psychology
Regression models
title Bayesian Active Learning for Discrete Latent Variable Models
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-14T13%3A18%3A29IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Bayesian%20Active%20Learning%20for%20Discrete%20Latent%20Variable%20Models&rft.jtitle=arXiv.org&rft.au=Jha,%20Aditi&rft.date=2023-06-02&rft.eissn=2331-8422&rft_id=info:doi/10.48550/arxiv.2202.13426&rft_dat=%3Cproquest%3E2634666157%3C/proquest%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-a956-bd7c284281d5dac14e8488554daaf48a51002653720ccf10278ef61d99d558683%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2634666157&rft_id=info:pmid/&rfr_iscdi=true