Loading…
Stochastic Gradient Descent for Kernel-Based Maximum Correntropy Criterion
Maximum correntropy criterion (MCC) has been an important method in machine learning and signal processing communities since it was successfully applied in various non-Gaussian noise scenarios. In comparison with the classical least squares method (LS), which takes only the second-order moment of mo...
Saved in:
Published in: | Entropy (Basel, Switzerland) Switzerland), 2024-12, Vol.26 (12), p.1104 |
---|---|
Main Authors: | , , , |
Format: | Article |
Language: | English |
Subjects: | |
Citations: | Items that this one cites |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
cited_by | |
---|---|
cites | cdi_FETCH-LOGICAL-c399t-4f4f1d8989858d770e99d88cb795541d78c6abd6d606a8efaa5b1327cc77ce023 |
container_end_page | |
container_issue | 12 |
container_start_page | 1104 |
container_title | Entropy (Basel, Switzerland) |
container_volume | 26 |
creator | Li, Tiankai Wang, Baobin Peng, Chaoquan Yin, Hong |
description | Maximum correntropy criterion (MCC) has been an important method in machine learning and signal processing communities since it was successfully applied in various non-Gaussian noise scenarios. In comparison with the classical least squares method (LS), which takes only the second-order moment of models into consideration and belongs to the convex optimization problem, MCC captures the high-order information of models that play crucial roles in robust learning, which is usually accompanied by solving the non-convexity optimization problems. As we know, the theoretical research on convex optimizations has made significant achievements, while theoretical understandings of non-convex optimization are still far from mature. Motivated by the popularity of the stochastic gradient descent (SGD) for solving nonconvex problems, this paper considers SGD applied to the kernel version of MCC, which has been shown to be robust to outliers and non-Gaussian data in nonlinear structure models. As the existing theoretical results for the SGD algorithm applied to the kernel MCC are not well established, we present the rigorous analysis for the convergence behaviors and provide explicit convergence rates under some standard conditions. Our work can fill the gap between optimization process and convergence during the iterations: the iterates need to converge to the global minimizer while the obtained estimator cannot ensure the global optimality in the learning process. |
doi_str_mv | 10.3390/e26121104 |
format | article |
fullrecord | <record><control><sourceid>gale_doaj_</sourceid><recordid>TN_cdi_doaj_primary_oai_doaj_org_article_f2e7e94d7da04bdf86643083ba7f3b27</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><galeid>A821766915</galeid><doaj_id>oai_doaj_org_article_f2e7e94d7da04bdf86643083ba7f3b27</doaj_id><sourcerecordid>A821766915</sourcerecordid><originalsourceid>FETCH-LOGICAL-c399t-4f4f1d8989858d770e99d88cb795541d78c6abd6d606a8efaa5b1327cc77ce023</originalsourceid><addsrcrecordid>eNpdkk1v1DAQhi0EomXhwB9AkbjQQ4odJ_44obKFUijiAJwtxx5vvUrirZ1F9N8zy5ZVi-Yw1vjxa7_jIeQlo6eca_oWGsEaxmj7iBwzqnXdckof31sfkWelrClteMPEU3LEtRRCcn5MPn-fk7u2ZY6uusjWR5jm6hyK2-WQcvUF8gRD_d4W8NVX-zuO27FappwRyGlzWy1znCHHND0nT4IdCry4ywvy8-OHH8tP9dW3i8vl2VXtuNZz3YY2MK80Rqe8lBS09kq5Xuqua5mXygnbe-EFFVZBsLbrGW-kc1I6QAsLcrnX9cmuzSbH0eZbk2w0fwspr4zN6GcAExqQoFsvvaVt74MSAtuheG9l4H0jUevdXmuz7UfwO9fZDg9EH-5M8dqs0i_DmJCdZi0qvLlTyOlmC2U2Y8TuDYOdIG2L4azjClmpEX39H7pO2zxhr5BqdScFbzhSp3tqZdFBnELCix2GhzG6NEGIWD9TDcM_1Ki-ICf7Ay6nUjKEw_MZNbv5MIf5QPbVfb8H8t9A8D94DbR1</addsrcrecordid><sourcetype>Open Website</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>3149576323</pqid></control><display><type>article</type><title>Stochastic Gradient Descent for Kernel-Based Maximum Correntropy Criterion</title><source>Publicly Available Content Database (Proquest) (PQ_SDU_P3)</source><source>DOAJ Directory of Open Access Journals</source><source>PubMed Central</source><creator>Li, Tiankai ; Wang, Baobin ; Peng, Chaoquan ; Yin, Hong</creator><creatorcontrib>Li, Tiankai ; Wang, Baobin ; Peng, Chaoquan ; Yin, Hong</creatorcontrib><description>Maximum correntropy criterion (MCC) has been an important method in machine learning and signal processing communities since it was successfully applied in various non-Gaussian noise scenarios. In comparison with the classical least squares method (LS), which takes only the second-order moment of models into consideration and belongs to the convex optimization problem, MCC captures the high-order information of models that play crucial roles in robust learning, which is usually accompanied by solving the non-convexity optimization problems. As we know, the theoretical research on convex optimizations has made significant achievements, while theoretical understandings of non-convex optimization are still far from mature. Motivated by the popularity of the stochastic gradient descent (SGD) for solving nonconvex problems, this paper considers SGD applied to the kernel version of MCC, which has been shown to be robust to outliers and non-Gaussian data in nonlinear structure models. As the existing theoretical results for the SGD algorithm applied to the kernel MCC are not well established, we present the rigorous analysis for the convergence behaviors and provide explicit convergence rates under some standard conditions. Our work can fill the gap between optimization process and convergence during the iterations: the iterates need to converge to the global minimizer while the obtained estimator cannot ensure the global optimality in the learning process.</description><identifier>ISSN: 1099-4300</identifier><identifier>EISSN: 1099-4300</identifier><identifier>DOI: 10.3390/e26121104</identifier><identifier>PMID: 39766733</identifier><language>eng</language><publisher>Switzerland: MDPI AG</publisher><subject>Algorithms ; Convergence ; convergence rate ; Convexity ; Criteria ; Data analysis ; Gaussian process ; Least squares method ; Machine learning ; maximum correntropy criterion ; non-Gaussian ; Optimization ; Outliers (statistics) ; Random noise ; Random variables ; Robustness ; Signal processing ; stochastic gradient descent</subject><ispartof>Entropy (Basel, Switzerland), 2024-12, Vol.26 (12), p.1104</ispartof><rights>COPYRIGHT 2024 MDPI AG</rights><rights>2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><rights>2024 by the authors. 2024</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><cites>FETCH-LOGICAL-c399t-4f4f1d8989858d770e99d88cb795541d78c6abd6d606a8efaa5b1327cc77ce023</cites><orcidid>0000-0003-0903-6344</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://www.proquest.com/docview/3149576323/fulltextPDF?pq-origsite=primo$$EPDF$$P50$$Gproquest$$Hfree_for_read</linktopdf><linktohtml>$$Uhttps://www.proquest.com/docview/3149576323?pq-origsite=primo$$EHTML$$P50$$Gproquest$$Hfree_for_read</linktohtml><link.rule.ids>230,314,723,776,780,860,881,2096,25731,27901,27902,36989,36990,44566,53766,53768,74869</link.rule.ids><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/39766733$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Li, Tiankai</creatorcontrib><creatorcontrib>Wang, Baobin</creatorcontrib><creatorcontrib>Peng, Chaoquan</creatorcontrib><creatorcontrib>Yin, Hong</creatorcontrib><title>Stochastic Gradient Descent for Kernel-Based Maximum Correntropy Criterion</title><title>Entropy (Basel, Switzerland)</title><addtitle>Entropy (Basel)</addtitle><description>Maximum correntropy criterion (MCC) has been an important method in machine learning and signal processing communities since it was successfully applied in various non-Gaussian noise scenarios. In comparison with the classical least squares method (LS), which takes only the second-order moment of models into consideration and belongs to the convex optimization problem, MCC captures the high-order information of models that play crucial roles in robust learning, which is usually accompanied by solving the non-convexity optimization problems. As we know, the theoretical research on convex optimizations has made significant achievements, while theoretical understandings of non-convex optimization are still far from mature. Motivated by the popularity of the stochastic gradient descent (SGD) for solving nonconvex problems, this paper considers SGD applied to the kernel version of MCC, which has been shown to be robust to outliers and non-Gaussian data in nonlinear structure models. As the existing theoretical results for the SGD algorithm applied to the kernel MCC are not well established, we present the rigorous analysis for the convergence behaviors and provide explicit convergence rates under some standard conditions. Our work can fill the gap between optimization process and convergence during the iterations: the iterates need to converge to the global minimizer while the obtained estimator cannot ensure the global optimality in the learning process.</description><subject>Algorithms</subject><subject>Convergence</subject><subject>convergence rate</subject><subject>Convexity</subject><subject>Criteria</subject><subject>Data analysis</subject><subject>Gaussian process</subject><subject>Least squares method</subject><subject>Machine learning</subject><subject>maximum correntropy criterion</subject><subject>non-Gaussian</subject><subject>Optimization</subject><subject>Outliers (statistics)</subject><subject>Random noise</subject><subject>Random variables</subject><subject>Robustness</subject><subject>Signal processing</subject><subject>stochastic gradient descent</subject><issn>1099-4300</issn><issn>1099-4300</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>PIMPY</sourceid><sourceid>DOA</sourceid><recordid>eNpdkk1v1DAQhi0EomXhwB9AkbjQQ4odJ_44obKFUijiAJwtxx5vvUrirZ1F9N8zy5ZVi-Yw1vjxa7_jIeQlo6eca_oWGsEaxmj7iBwzqnXdckof31sfkWelrClteMPEU3LEtRRCcn5MPn-fk7u2ZY6uusjWR5jm6hyK2-WQcvUF8gRD_d4W8NVX-zuO27FappwRyGlzWy1znCHHND0nT4IdCry4ywvy8-OHH8tP9dW3i8vl2VXtuNZz3YY2MK80Rqe8lBS09kq5Xuqua5mXygnbe-EFFVZBsLbrGW-kc1I6QAsLcrnX9cmuzSbH0eZbk2w0fwspr4zN6GcAExqQoFsvvaVt74MSAtuheG9l4H0jUevdXmuz7UfwO9fZDg9EH-5M8dqs0i_DmJCdZi0qvLlTyOlmC2U2Y8TuDYOdIG2L4azjClmpEX39H7pO2zxhr5BqdScFbzhSp3tqZdFBnELCix2GhzG6NEGIWD9TDcM_1Ki-ICf7Ay6nUjKEw_MZNbv5MIf5QPbVfb8H8t9A8D94DbR1</recordid><startdate>20241217</startdate><enddate>20241217</enddate><creator>Li, Tiankai</creator><creator>Wang, Baobin</creator><creator>Peng, Chaoquan</creator><creator>Yin, Hong</creator><general>MDPI AG</general><general>MDPI</general><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7TB</scope><scope>8FD</scope><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>FR3</scope><scope>HCIFZ</scope><scope>KR7</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope><scope>7X8</scope><scope>5PM</scope><scope>DOA</scope><orcidid>https://orcid.org/0000-0003-0903-6344</orcidid></search><sort><creationdate>20241217</creationdate><title>Stochastic Gradient Descent for Kernel-Based Maximum Correntropy Criterion</title><author>Li, Tiankai ; Wang, Baobin ; Peng, Chaoquan ; Yin, Hong</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c399t-4f4f1d8989858d770e99d88cb795541d78c6abd6d606a8efaa5b1327cc77ce023</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Algorithms</topic><topic>Convergence</topic><topic>convergence rate</topic><topic>Convexity</topic><topic>Criteria</topic><topic>Data analysis</topic><topic>Gaussian process</topic><topic>Least squares method</topic><topic>Machine learning</topic><topic>maximum correntropy criterion</topic><topic>non-Gaussian</topic><topic>Optimization</topic><topic>Outliers (statistics)</topic><topic>Random noise</topic><topic>Random variables</topic><topic>Robustness</topic><topic>Signal processing</topic><topic>stochastic gradient descent</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Li, Tiankai</creatorcontrib><creatorcontrib>Wang, Baobin</creatorcontrib><creatorcontrib>Peng, Chaoquan</creatorcontrib><creatorcontrib>Yin, Hong</creatorcontrib><collection>PubMed</collection><collection>CrossRef</collection><collection>Mechanical & Transportation Engineering Abstracts</collection><collection>Technology Research Database</collection><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science & Engineering Collection</collection><collection>ProQuest Central (Alumni)</collection><collection>ProQuest Central</collection><collection>ProQuest Central Essentials</collection><collection>AUTh Library subscriptions: ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central</collection><collection>Engineering Research Database</collection><collection>SciTech Premium Collection</collection><collection>Civil Engineering Abstracts</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Publicly Available Content Database (Proquest) (PQ_SDU_P3)</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering collection</collection><collection>MEDLINE - Academic</collection><collection>PubMed Central (Full Participant titles)</collection><collection>DOAJ Directory of Open Access Journals</collection><jtitle>Entropy (Basel, Switzerland)</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Li, Tiankai</au><au>Wang, Baobin</au><au>Peng, Chaoquan</au><au>Yin, Hong</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Stochastic Gradient Descent for Kernel-Based Maximum Correntropy Criterion</atitle><jtitle>Entropy (Basel, Switzerland)</jtitle><addtitle>Entropy (Basel)</addtitle><date>2024-12-17</date><risdate>2024</risdate><volume>26</volume><issue>12</issue><spage>1104</spage><pages>1104-</pages><issn>1099-4300</issn><eissn>1099-4300</eissn><abstract>Maximum correntropy criterion (MCC) has been an important method in machine learning and signal processing communities since it was successfully applied in various non-Gaussian noise scenarios. In comparison with the classical least squares method (LS), which takes only the second-order moment of models into consideration and belongs to the convex optimization problem, MCC captures the high-order information of models that play crucial roles in robust learning, which is usually accompanied by solving the non-convexity optimization problems. As we know, the theoretical research on convex optimizations has made significant achievements, while theoretical understandings of non-convex optimization are still far from mature. Motivated by the popularity of the stochastic gradient descent (SGD) for solving nonconvex problems, this paper considers SGD applied to the kernel version of MCC, which has been shown to be robust to outliers and non-Gaussian data in nonlinear structure models. As the existing theoretical results for the SGD algorithm applied to the kernel MCC are not well established, we present the rigorous analysis for the convergence behaviors and provide explicit convergence rates under some standard conditions. Our work can fill the gap between optimization process and convergence during the iterations: the iterates need to converge to the global minimizer while the obtained estimator cannot ensure the global optimality in the learning process.</abstract><cop>Switzerland</cop><pub>MDPI AG</pub><pmid>39766733</pmid><doi>10.3390/e26121104</doi><orcidid>https://orcid.org/0000-0003-0903-6344</orcidid><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | ISSN: 1099-4300 |
ispartof | Entropy (Basel, Switzerland), 2024-12, Vol.26 (12), p.1104 |
issn | 1099-4300 1099-4300 |
language | eng |
recordid | cdi_doaj_primary_oai_doaj_org_article_f2e7e94d7da04bdf86643083ba7f3b27 |
source | Publicly Available Content Database (Proquest) (PQ_SDU_P3); DOAJ Directory of Open Access Journals; PubMed Central |
subjects | Algorithms Convergence convergence rate Convexity Criteria Data analysis Gaussian process Least squares method Machine learning maximum correntropy criterion non-Gaussian Optimization Outliers (statistics) Random noise Random variables Robustness Signal processing stochastic gradient descent |
title | Stochastic Gradient Descent for Kernel-Based Maximum Correntropy Criterion |
url | http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-10T13%3A21%3A40IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-gale_doaj_&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Stochastic%20Gradient%20Descent%20for%20Kernel-Based%20Maximum%20Correntropy%20Criterion&rft.jtitle=Entropy%20(Basel,%20Switzerland)&rft.au=Li,%20Tiankai&rft.date=2024-12-17&rft.volume=26&rft.issue=12&rft.spage=1104&rft.pages=1104-&rft.issn=1099-4300&rft.eissn=1099-4300&rft_id=info:doi/10.3390/e26121104&rft_dat=%3Cgale_doaj_%3EA821766915%3C/gale_doaj_%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c399t-4f4f1d8989858d770e99d88cb795541d78c6abd6d606a8efaa5b1327cc77ce023%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=3149576323&rft_id=info:pmid/39766733&rft_galeid=A821766915&rfr_iscdi=true |