Loading…

Stochastic Gradient Descent for Kernel-Based Maximum Correntropy Criterion

Maximum correntropy criterion (MCC) has been an important method in machine learning and signal processing communities since it was successfully applied in various non-Gaussian noise scenarios. In comparison with the classical least squares method (LS), which takes only the second-order moment of mo...

Full description

Saved in:
Bibliographic Details
Published in:Entropy (Basel, Switzerland) Switzerland), 2024-12, Vol.26 (12), p.1104
Main Authors: Li, Tiankai, Wang, Baobin, Peng, Chaoquan, Yin, Hong
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by
cites cdi_FETCH-LOGICAL-c399t-4f4f1d8989858d770e99d88cb795541d78c6abd6d606a8efaa5b1327cc77ce023
container_end_page
container_issue 12
container_start_page 1104
container_title Entropy (Basel, Switzerland)
container_volume 26
creator Li, Tiankai
Wang, Baobin
Peng, Chaoquan
Yin, Hong
description Maximum correntropy criterion (MCC) has been an important method in machine learning and signal processing communities since it was successfully applied in various non-Gaussian noise scenarios. In comparison with the classical least squares method (LS), which takes only the second-order moment of models into consideration and belongs to the convex optimization problem, MCC captures the high-order information of models that play crucial roles in robust learning, which is usually accompanied by solving the non-convexity optimization problems. As we know, the theoretical research on convex optimizations has made significant achievements, while theoretical understandings of non-convex optimization are still far from mature. Motivated by the popularity of the stochastic gradient descent (SGD) for solving nonconvex problems, this paper considers SGD applied to the kernel version of MCC, which has been shown to be robust to outliers and non-Gaussian data in nonlinear structure models. As the existing theoretical results for the SGD algorithm applied to the kernel MCC are not well established, we present the rigorous analysis for the convergence behaviors and provide explicit convergence rates under some standard conditions. Our work can fill the gap between optimization process and convergence during the iterations: the iterates need to converge to the global minimizer while the obtained estimator cannot ensure the global optimality in the learning process.
doi_str_mv 10.3390/e26121104
format article
fullrecord <record><control><sourceid>gale_doaj_</sourceid><recordid>TN_cdi_doaj_primary_oai_doaj_org_article_f2e7e94d7da04bdf86643083ba7f3b27</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><galeid>A821766915</galeid><doaj_id>oai_doaj_org_article_f2e7e94d7da04bdf86643083ba7f3b27</doaj_id><sourcerecordid>A821766915</sourcerecordid><originalsourceid>FETCH-LOGICAL-c399t-4f4f1d8989858d770e99d88cb795541d78c6abd6d606a8efaa5b1327cc77ce023</originalsourceid><addsrcrecordid>eNpdkk1v1DAQhi0EomXhwB9AkbjQQ4odJ_44obKFUijiAJwtxx5vvUrirZ1F9N8zy5ZVi-Yw1vjxa7_jIeQlo6eca_oWGsEaxmj7iBwzqnXdckof31sfkWelrClteMPEU3LEtRRCcn5MPn-fk7u2ZY6uusjWR5jm6hyK2-WQcvUF8gRD_d4W8NVX-zuO27FappwRyGlzWy1znCHHND0nT4IdCry4ywvy8-OHH8tP9dW3i8vl2VXtuNZz3YY2MK80Rqe8lBS09kq5Xuqua5mXygnbe-EFFVZBsLbrGW-kc1I6QAsLcrnX9cmuzSbH0eZbk2w0fwspr4zN6GcAExqQoFsvvaVt74MSAtuheG9l4H0jUevdXmuz7UfwO9fZDg9EH-5M8dqs0i_DmJCdZi0qvLlTyOlmC2U2Y8TuDYOdIG2L4azjClmpEX39H7pO2zxhr5BqdScFbzhSp3tqZdFBnELCix2GhzG6NEGIWD9TDcM_1Ki-ICf7Ay6nUjKEw_MZNbv5MIf5QPbVfb8H8t9A8D94DbR1</addsrcrecordid><sourcetype>Open Website</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>3149576323</pqid></control><display><type>article</type><title>Stochastic Gradient Descent for Kernel-Based Maximum Correntropy Criterion</title><source>Publicly Available Content Database (Proquest) (PQ_SDU_P3)</source><source>DOAJ Directory of Open Access Journals</source><source>PubMed Central</source><creator>Li, Tiankai ; Wang, Baobin ; Peng, Chaoquan ; Yin, Hong</creator><creatorcontrib>Li, Tiankai ; Wang, Baobin ; Peng, Chaoquan ; Yin, Hong</creatorcontrib><description>Maximum correntropy criterion (MCC) has been an important method in machine learning and signal processing communities since it was successfully applied in various non-Gaussian noise scenarios. In comparison with the classical least squares method (LS), which takes only the second-order moment of models into consideration and belongs to the convex optimization problem, MCC captures the high-order information of models that play crucial roles in robust learning, which is usually accompanied by solving the non-convexity optimization problems. As we know, the theoretical research on convex optimizations has made significant achievements, while theoretical understandings of non-convex optimization are still far from mature. Motivated by the popularity of the stochastic gradient descent (SGD) for solving nonconvex problems, this paper considers SGD applied to the kernel version of MCC, which has been shown to be robust to outliers and non-Gaussian data in nonlinear structure models. As the existing theoretical results for the SGD algorithm applied to the kernel MCC are not well established, we present the rigorous analysis for the convergence behaviors and provide explicit convergence rates under some standard conditions. Our work can fill the gap between optimization process and convergence during the iterations: the iterates need to converge to the global minimizer while the obtained estimator cannot ensure the global optimality in the learning process.</description><identifier>ISSN: 1099-4300</identifier><identifier>EISSN: 1099-4300</identifier><identifier>DOI: 10.3390/e26121104</identifier><identifier>PMID: 39766733</identifier><language>eng</language><publisher>Switzerland: MDPI AG</publisher><subject>Algorithms ; Convergence ; convergence rate ; Convexity ; Criteria ; Data analysis ; Gaussian process ; Least squares method ; Machine learning ; maximum correntropy criterion ; non-Gaussian ; Optimization ; Outliers (statistics) ; Random noise ; Random variables ; Robustness ; Signal processing ; stochastic gradient descent</subject><ispartof>Entropy (Basel, Switzerland), 2024-12, Vol.26 (12), p.1104</ispartof><rights>COPYRIGHT 2024 MDPI AG</rights><rights>2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><rights>2024 by the authors. 2024</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><cites>FETCH-LOGICAL-c399t-4f4f1d8989858d770e99d88cb795541d78c6abd6d606a8efaa5b1327cc77ce023</cites><orcidid>0000-0003-0903-6344</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://www.proquest.com/docview/3149576323/fulltextPDF?pq-origsite=primo$$EPDF$$P50$$Gproquest$$Hfree_for_read</linktopdf><linktohtml>$$Uhttps://www.proquest.com/docview/3149576323?pq-origsite=primo$$EHTML$$P50$$Gproquest$$Hfree_for_read</linktohtml><link.rule.ids>230,314,723,776,780,860,881,2096,25731,27901,27902,36989,36990,44566,53766,53768,74869</link.rule.ids><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/39766733$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Li, Tiankai</creatorcontrib><creatorcontrib>Wang, Baobin</creatorcontrib><creatorcontrib>Peng, Chaoquan</creatorcontrib><creatorcontrib>Yin, Hong</creatorcontrib><title>Stochastic Gradient Descent for Kernel-Based Maximum Correntropy Criterion</title><title>Entropy (Basel, Switzerland)</title><addtitle>Entropy (Basel)</addtitle><description>Maximum correntropy criterion (MCC) has been an important method in machine learning and signal processing communities since it was successfully applied in various non-Gaussian noise scenarios. In comparison with the classical least squares method (LS), which takes only the second-order moment of models into consideration and belongs to the convex optimization problem, MCC captures the high-order information of models that play crucial roles in robust learning, which is usually accompanied by solving the non-convexity optimization problems. As we know, the theoretical research on convex optimizations has made significant achievements, while theoretical understandings of non-convex optimization are still far from mature. Motivated by the popularity of the stochastic gradient descent (SGD) for solving nonconvex problems, this paper considers SGD applied to the kernel version of MCC, which has been shown to be robust to outliers and non-Gaussian data in nonlinear structure models. As the existing theoretical results for the SGD algorithm applied to the kernel MCC are not well established, we present the rigorous analysis for the convergence behaviors and provide explicit convergence rates under some standard conditions. Our work can fill the gap between optimization process and convergence during the iterations: the iterates need to converge to the global minimizer while the obtained estimator cannot ensure the global optimality in the learning process.</description><subject>Algorithms</subject><subject>Convergence</subject><subject>convergence rate</subject><subject>Convexity</subject><subject>Criteria</subject><subject>Data analysis</subject><subject>Gaussian process</subject><subject>Least squares method</subject><subject>Machine learning</subject><subject>maximum correntropy criterion</subject><subject>non-Gaussian</subject><subject>Optimization</subject><subject>Outliers (statistics)</subject><subject>Random noise</subject><subject>Random variables</subject><subject>Robustness</subject><subject>Signal processing</subject><subject>stochastic gradient descent</subject><issn>1099-4300</issn><issn>1099-4300</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>PIMPY</sourceid><sourceid>DOA</sourceid><recordid>eNpdkk1v1DAQhi0EomXhwB9AkbjQQ4odJ_44obKFUijiAJwtxx5vvUrirZ1F9N8zy5ZVi-Yw1vjxa7_jIeQlo6eca_oWGsEaxmj7iBwzqnXdckof31sfkWelrClteMPEU3LEtRRCcn5MPn-fk7u2ZY6uusjWR5jm6hyK2-WQcvUF8gRD_d4W8NVX-zuO27FappwRyGlzWy1znCHHND0nT4IdCry4ywvy8-OHH8tP9dW3i8vl2VXtuNZz3YY2MK80Rqe8lBS09kq5Xuqua5mXygnbe-EFFVZBsLbrGW-kc1I6QAsLcrnX9cmuzSbH0eZbk2w0fwspr4zN6GcAExqQoFsvvaVt74MSAtuheG9l4H0jUevdXmuz7UfwO9fZDg9EH-5M8dqs0i_DmJCdZi0qvLlTyOlmC2U2Y8TuDYOdIG2L4azjClmpEX39H7pO2zxhr5BqdScFbzhSp3tqZdFBnELCix2GhzG6NEGIWD9TDcM_1Ki-ICf7Ay6nUjKEw_MZNbv5MIf5QPbVfb8H8t9A8D94DbR1</recordid><startdate>20241217</startdate><enddate>20241217</enddate><creator>Li, Tiankai</creator><creator>Wang, Baobin</creator><creator>Peng, Chaoquan</creator><creator>Yin, Hong</creator><general>MDPI AG</general><general>MDPI</general><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7TB</scope><scope>8FD</scope><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>FR3</scope><scope>HCIFZ</scope><scope>KR7</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope><scope>7X8</scope><scope>5PM</scope><scope>DOA</scope><orcidid>https://orcid.org/0000-0003-0903-6344</orcidid></search><sort><creationdate>20241217</creationdate><title>Stochastic Gradient Descent for Kernel-Based Maximum Correntropy Criterion</title><author>Li, Tiankai ; Wang, Baobin ; Peng, Chaoquan ; Yin, Hong</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c399t-4f4f1d8989858d770e99d88cb795541d78c6abd6d606a8efaa5b1327cc77ce023</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Algorithms</topic><topic>Convergence</topic><topic>convergence rate</topic><topic>Convexity</topic><topic>Criteria</topic><topic>Data analysis</topic><topic>Gaussian process</topic><topic>Least squares method</topic><topic>Machine learning</topic><topic>maximum correntropy criterion</topic><topic>non-Gaussian</topic><topic>Optimization</topic><topic>Outliers (statistics)</topic><topic>Random noise</topic><topic>Random variables</topic><topic>Robustness</topic><topic>Signal processing</topic><topic>stochastic gradient descent</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Li, Tiankai</creatorcontrib><creatorcontrib>Wang, Baobin</creatorcontrib><creatorcontrib>Peng, Chaoquan</creatorcontrib><creatorcontrib>Yin, Hong</creatorcontrib><collection>PubMed</collection><collection>CrossRef</collection><collection>Mechanical &amp; Transportation Engineering Abstracts</collection><collection>Technology Research Database</collection><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science &amp; Engineering Collection</collection><collection>ProQuest Central (Alumni)</collection><collection>ProQuest Central</collection><collection>ProQuest Central Essentials</collection><collection>AUTh Library subscriptions: ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central</collection><collection>Engineering Research Database</collection><collection>SciTech Premium Collection</collection><collection>Civil Engineering Abstracts</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Publicly Available Content Database (Proquest) (PQ_SDU_P3)</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering collection</collection><collection>MEDLINE - Academic</collection><collection>PubMed Central (Full Participant titles)</collection><collection>DOAJ Directory of Open Access Journals</collection><jtitle>Entropy (Basel, Switzerland)</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Li, Tiankai</au><au>Wang, Baobin</au><au>Peng, Chaoquan</au><au>Yin, Hong</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Stochastic Gradient Descent for Kernel-Based Maximum Correntropy Criterion</atitle><jtitle>Entropy (Basel, Switzerland)</jtitle><addtitle>Entropy (Basel)</addtitle><date>2024-12-17</date><risdate>2024</risdate><volume>26</volume><issue>12</issue><spage>1104</spage><pages>1104-</pages><issn>1099-4300</issn><eissn>1099-4300</eissn><abstract>Maximum correntropy criterion (MCC) has been an important method in machine learning and signal processing communities since it was successfully applied in various non-Gaussian noise scenarios. In comparison with the classical least squares method (LS), which takes only the second-order moment of models into consideration and belongs to the convex optimization problem, MCC captures the high-order information of models that play crucial roles in robust learning, which is usually accompanied by solving the non-convexity optimization problems. As we know, the theoretical research on convex optimizations has made significant achievements, while theoretical understandings of non-convex optimization are still far from mature. Motivated by the popularity of the stochastic gradient descent (SGD) for solving nonconvex problems, this paper considers SGD applied to the kernel version of MCC, which has been shown to be robust to outliers and non-Gaussian data in nonlinear structure models. As the existing theoretical results for the SGD algorithm applied to the kernel MCC are not well established, we present the rigorous analysis for the convergence behaviors and provide explicit convergence rates under some standard conditions. Our work can fill the gap between optimization process and convergence during the iterations: the iterates need to converge to the global minimizer while the obtained estimator cannot ensure the global optimality in the learning process.</abstract><cop>Switzerland</cop><pub>MDPI AG</pub><pmid>39766733</pmid><doi>10.3390/e26121104</doi><orcidid>https://orcid.org/0000-0003-0903-6344</orcidid><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 1099-4300
ispartof Entropy (Basel, Switzerland), 2024-12, Vol.26 (12), p.1104
issn 1099-4300
1099-4300
language eng
recordid cdi_doaj_primary_oai_doaj_org_article_f2e7e94d7da04bdf86643083ba7f3b27
source Publicly Available Content Database (Proquest) (PQ_SDU_P3); DOAJ Directory of Open Access Journals; PubMed Central
subjects Algorithms
Convergence
convergence rate
Convexity
Criteria
Data analysis
Gaussian process
Least squares method
Machine learning
maximum correntropy criterion
non-Gaussian
Optimization
Outliers (statistics)
Random noise
Random variables
Robustness
Signal processing
stochastic gradient descent
title Stochastic Gradient Descent for Kernel-Based Maximum Correntropy Criterion
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-10T13%3A21%3A40IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-gale_doaj_&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Stochastic%20Gradient%20Descent%20for%20Kernel-Based%20Maximum%20Correntropy%20Criterion&rft.jtitle=Entropy%20(Basel,%20Switzerland)&rft.au=Li,%20Tiankai&rft.date=2024-12-17&rft.volume=26&rft.issue=12&rft.spage=1104&rft.pages=1104-&rft.issn=1099-4300&rft.eissn=1099-4300&rft_id=info:doi/10.3390/e26121104&rft_dat=%3Cgale_doaj_%3EA821766915%3C/gale_doaj_%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c399t-4f4f1d8989858d770e99d88cb795541d78c6abd6d606a8efaa5b1327cc77ce023%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=3149576323&rft_id=info:pmid/39766733&rft_galeid=A821766915&rfr_iscdi=true