Loading…

Multi-Task Learning with Calibrated Mixture of Insightful Experts

Multi-task learning has been established as an important machine learning framework for leveraging shared knowledge among multiple different but related tasks, with the generalization performance of models enhanced. As a promising learning paradigm, multi-task learning has been widely adopted by var...

Full description

Saved in:
Bibliographic Details
Main Authors: Wang, Sinan, Li, Yumeng, Li, Hongyan, Zhu, Tanchao, Li, Zhao, Ou, Wenwu
Format: Conference Proceeding
Language:English
Subjects:
Online Access:Request full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by
cites
container_end_page 3319
container_issue
container_start_page 3307
container_title
container_volume
creator Wang, Sinan
Li, Yumeng
Li, Hongyan
Zhu, Tanchao
Li, Zhao
Ou, Wenwu
description Multi-task learning has been established as an important machine learning framework for leveraging shared knowledge among multiple different but related tasks, with the generalization performance of models enhanced. As a promising learning paradigm, multi-task learning has been widely adopted by various real-world applications, such as recommendation systems. Multi-gate Mixture-of-Experts (MMoE), a well-received multi-task learning method in industry, based on the classic and inspiring Mixture-of-Experts (MoE) structure, explicitly models task relationships and learns task-specific functionalities, generating significant improvements. However, in our applications, negative transfer, which confuses considerable existing multi-task learning methods, is still observed to happen to MMoE. In this paper, an in-depth empirical investigation into negative transfer is launched. And it reveals that, incompetent experts, which play fundamental roles under the learning framework of MoE, are the key technique bottleneck. To tackle this dilemma, we propose the Calibrated Mixture of Insightful Experts (CMoIE), with three novel modules (Conflict Resolution, Expert Communication, and Mixture Calibration), customed for multi-task learning. Hence a group of insightful experts are constructed with enhanced diversity, communication and specialization. To validate the proposed method CMoIE, experiments are conducted on three public datasets and one real-world click-through-rate prediction dataset we construct based on traffic logs collected from a large-scale online product recommendation system. Our approach yields best performance across all of these benchmarks, demonstrating the superiority of it.
doi_str_mv 10.1109/ICDE53745.2022.00312
format conference_proceeding
fullrecord <record><control><sourceid>ieee_CHZPO</sourceid><recordid>TN_cdi_ieee_primary_9835373</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>9835373</ieee_id><sourcerecordid>9835373</sourcerecordid><originalsourceid>FETCH-LOGICAL-i203t-e8d52adc140406658ab798dc4deb8ecd7aac026ef7cf152a1ab13f4bb0c5cbab3</originalsourceid><addsrcrecordid>eNotj8tKw0AYhUdBsNY-gS7mBRL_uSWTZYmxBlLcVHBX5pZ2NMYyM8H69gb0bM7m8PEdhO4J5IRA9dDWj41gJRc5BUpzAEboBbohRSE4SMmqS7SgrBQZ0OLtGq1ifIc5FSdEwAKtt9OQfLZT8QN3ToXRjwf87dMR12rwOqjkLN76c5qCw189bsfoD8fUTwNuzicXUrxFV70aolv99xK9PjW7-jnrXjZtve4yT4GlzEkrqLKGcOAwy0mly0paw63T0hlbKmVmRdeXpifzkihNWM-1BiOMVpot0d0f1zvn9qfgP1X42VeSze8Z-wUAp0vg</addsrcrecordid><sourcetype>Publisher</sourcetype><iscdi>true</iscdi><recordtype>conference_proceeding</recordtype></control><display><type>conference_proceeding</type><title>Multi-Task Learning with Calibrated Mixture of Insightful Experts</title><source>IEEE Xplore All Conference Series</source><creator>Wang, Sinan ; Li, Yumeng ; Li, Hongyan ; Zhu, Tanchao ; Li, Zhao ; Ou, Wenwu</creator><creatorcontrib>Wang, Sinan ; Li, Yumeng ; Li, Hongyan ; Zhu, Tanchao ; Li, Zhao ; Ou, Wenwu</creatorcontrib><description>Multi-task learning has been established as an important machine learning framework for leveraging shared knowledge among multiple different but related tasks, with the generalization performance of models enhanced. As a promising learning paradigm, multi-task learning has been widely adopted by various real-world applications, such as recommendation systems. Multi-gate Mixture-of-Experts (MMoE), a well-received multi-task learning method in industry, based on the classic and inspiring Mixture-of-Experts (MoE) structure, explicitly models task relationships and learns task-specific functionalities, generating significant improvements. However, in our applications, negative transfer, which confuses considerable existing multi-task learning methods, is still observed to happen to MMoE. In this paper, an in-depth empirical investigation into negative transfer is launched. And it reveals that, incompetent experts, which play fundamental roles under the learning framework of MoE, are the key technique bottleneck. To tackle this dilemma, we propose the Calibrated Mixture of Insightful Experts (CMoIE), with three novel modules (Conflict Resolution, Expert Communication, and Mixture Calibration), customed for multi-task learning. Hence a group of insightful experts are constructed with enhanced diversity, communication and specialization. To validate the proposed method CMoIE, experiments are conducted on three public datasets and one real-world click-through-rate prediction dataset we construct based on traffic logs collected from a large-scale online product recommendation system. Our approach yields best performance across all of these benchmarks, demonstrating the superiority of it.</description><identifier>EISSN: 2375-026X</identifier><identifier>EISBN: 1665408839</identifier><identifier>EISBN: 9781665408837</identifier><identifier>DOI: 10.1109/ICDE53745.2022.00312</identifier><identifier>CODEN: IEEPAD</identifier><language>eng</language><publisher>IEEE</publisher><subject>Benchmark testing ; Conferences ; Industries ; Learning systems ; Machine learning ; mixture-of-experts ; multi-task learning ; Multitasking ; Network architecture ; recommendation systems</subject><ispartof>2022 IEEE 38th International Conference on Data Engineering (ICDE), 2022, p.3307-3319</ispartof><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/9835373$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>309,310,780,784,789,790,23930,23931,25140,27925,54555,54932</link.rule.ids><linktorsrc>$$Uhttps://ieeexplore.ieee.org/document/9835373$$EView_record_in_IEEE$$FView_record_in_$$GIEEE</linktorsrc></links><search><creatorcontrib>Wang, Sinan</creatorcontrib><creatorcontrib>Li, Yumeng</creatorcontrib><creatorcontrib>Li, Hongyan</creatorcontrib><creatorcontrib>Zhu, Tanchao</creatorcontrib><creatorcontrib>Li, Zhao</creatorcontrib><creatorcontrib>Ou, Wenwu</creatorcontrib><title>Multi-Task Learning with Calibrated Mixture of Insightful Experts</title><title>2022 IEEE 38th International Conference on Data Engineering (ICDE)</title><addtitle>ICDE</addtitle><description>Multi-task learning has been established as an important machine learning framework for leveraging shared knowledge among multiple different but related tasks, with the generalization performance of models enhanced. As a promising learning paradigm, multi-task learning has been widely adopted by various real-world applications, such as recommendation systems. Multi-gate Mixture-of-Experts (MMoE), a well-received multi-task learning method in industry, based on the classic and inspiring Mixture-of-Experts (MoE) structure, explicitly models task relationships and learns task-specific functionalities, generating significant improvements. However, in our applications, negative transfer, which confuses considerable existing multi-task learning methods, is still observed to happen to MMoE. In this paper, an in-depth empirical investigation into negative transfer is launched. And it reveals that, incompetent experts, which play fundamental roles under the learning framework of MoE, are the key technique bottleneck. To tackle this dilemma, we propose the Calibrated Mixture of Insightful Experts (CMoIE), with three novel modules (Conflict Resolution, Expert Communication, and Mixture Calibration), customed for multi-task learning. Hence a group of insightful experts are constructed with enhanced diversity, communication and specialization. To validate the proposed method CMoIE, experiments are conducted on three public datasets and one real-world click-through-rate prediction dataset we construct based on traffic logs collected from a large-scale online product recommendation system. Our approach yields best performance across all of these benchmarks, demonstrating the superiority of it.</description><subject>Benchmark testing</subject><subject>Conferences</subject><subject>Industries</subject><subject>Learning systems</subject><subject>Machine learning</subject><subject>mixture-of-experts</subject><subject>multi-task learning</subject><subject>Multitasking</subject><subject>Network architecture</subject><subject>recommendation systems</subject><issn>2375-026X</issn><isbn>1665408839</isbn><isbn>9781665408837</isbn><fulltext>true</fulltext><rsrctype>conference_proceeding</rsrctype><creationdate>2022</creationdate><recordtype>conference_proceeding</recordtype><sourceid>6IE</sourceid><recordid>eNotj8tKw0AYhUdBsNY-gS7mBRL_uSWTZYmxBlLcVHBX5pZ2NMYyM8H69gb0bM7m8PEdhO4J5IRA9dDWj41gJRc5BUpzAEboBbohRSE4SMmqS7SgrBQZ0OLtGq1ifIc5FSdEwAKtt9OQfLZT8QN3ToXRjwf87dMR12rwOqjkLN76c5qCw189bsfoD8fUTwNuzicXUrxFV70aolv99xK9PjW7-jnrXjZtve4yT4GlzEkrqLKGcOAwy0mly0paw63T0hlbKmVmRdeXpifzkihNWM-1BiOMVpot0d0f1zvn9qfgP1X42VeSze8Z-wUAp0vg</recordid><startdate>202205</startdate><enddate>202205</enddate><creator>Wang, Sinan</creator><creator>Li, Yumeng</creator><creator>Li, Hongyan</creator><creator>Zhu, Tanchao</creator><creator>Li, Zhao</creator><creator>Ou, Wenwu</creator><general>IEEE</general><scope>6IE</scope><scope>6IH</scope><scope>CBEJK</scope><scope>RIE</scope><scope>RIO</scope></search><sort><creationdate>202205</creationdate><title>Multi-Task Learning with Calibrated Mixture of Insightful Experts</title><author>Wang, Sinan ; Li, Yumeng ; Li, Hongyan ; Zhu, Tanchao ; Li, Zhao ; Ou, Wenwu</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-i203t-e8d52adc140406658ab798dc4deb8ecd7aac026ef7cf152a1ab13f4bb0c5cbab3</frbrgroupid><rsrctype>conference_proceedings</rsrctype><prefilter>conference_proceedings</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Benchmark testing</topic><topic>Conferences</topic><topic>Industries</topic><topic>Learning systems</topic><topic>Machine learning</topic><topic>mixture-of-experts</topic><topic>multi-task learning</topic><topic>Multitasking</topic><topic>Network architecture</topic><topic>recommendation systems</topic><toplevel>online_resources</toplevel><creatorcontrib>Wang, Sinan</creatorcontrib><creatorcontrib>Li, Yumeng</creatorcontrib><creatorcontrib>Li, Hongyan</creatorcontrib><creatorcontrib>Zhu, Tanchao</creatorcontrib><creatorcontrib>Li, Zhao</creatorcontrib><creatorcontrib>Ou, Wenwu</creatorcontrib><collection>IEEE Electronic Library (IEL) Conference Proceedings</collection><collection>IEEE Proceedings Order Plan (POP) 1998-present by volume</collection><collection>IEEE Xplore All Conference Proceedings</collection><collection>IEL</collection><collection>IEEE Proceedings Order Plans (POP) 1998-present</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Wang, Sinan</au><au>Li, Yumeng</au><au>Li, Hongyan</au><au>Zhu, Tanchao</au><au>Li, Zhao</au><au>Ou, Wenwu</au><format>book</format><genre>proceeding</genre><ristype>CONF</ristype><atitle>Multi-Task Learning with Calibrated Mixture of Insightful Experts</atitle><btitle>2022 IEEE 38th International Conference on Data Engineering (ICDE)</btitle><stitle>ICDE</stitle><date>2022-05</date><risdate>2022</risdate><spage>3307</spage><epage>3319</epage><pages>3307-3319</pages><eissn>2375-026X</eissn><eisbn>1665408839</eisbn><eisbn>9781665408837</eisbn><coden>IEEPAD</coden><abstract>Multi-task learning has been established as an important machine learning framework for leveraging shared knowledge among multiple different but related tasks, with the generalization performance of models enhanced. As a promising learning paradigm, multi-task learning has been widely adopted by various real-world applications, such as recommendation systems. Multi-gate Mixture-of-Experts (MMoE), a well-received multi-task learning method in industry, based on the classic and inspiring Mixture-of-Experts (MoE) structure, explicitly models task relationships and learns task-specific functionalities, generating significant improvements. However, in our applications, negative transfer, which confuses considerable existing multi-task learning methods, is still observed to happen to MMoE. In this paper, an in-depth empirical investigation into negative transfer is launched. And it reveals that, incompetent experts, which play fundamental roles under the learning framework of MoE, are the key technique bottleneck. To tackle this dilemma, we propose the Calibrated Mixture of Insightful Experts (CMoIE), with three novel modules (Conflict Resolution, Expert Communication, and Mixture Calibration), customed for multi-task learning. Hence a group of insightful experts are constructed with enhanced diversity, communication and specialization. To validate the proposed method CMoIE, experiments are conducted on three public datasets and one real-world click-through-rate prediction dataset we construct based on traffic logs collected from a large-scale online product recommendation system. Our approach yields best performance across all of these benchmarks, demonstrating the superiority of it.</abstract><pub>IEEE</pub><doi>10.1109/ICDE53745.2022.00312</doi><tpages>13</tpages></addata></record>
fulltext fulltext_linktorsrc
identifier EISSN: 2375-026X
ispartof 2022 IEEE 38th International Conference on Data Engineering (ICDE), 2022, p.3307-3319
issn 2375-026X
language eng
recordid cdi_ieee_primary_9835373
source IEEE Xplore All Conference Series
subjects Benchmark testing
Conferences
Industries
Learning systems
Machine learning
mixture-of-experts
multi-task learning
Multitasking
Network architecture
recommendation systems
title Multi-Task Learning with Calibrated Mixture of Insightful Experts
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-02T10%3A30%3A56IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-ieee_CHZPO&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=proceeding&rft.atitle=Multi-Task%20Learning%20with%20Calibrated%20Mixture%20of%20Insightful%20Experts&rft.btitle=2022%20IEEE%2038th%20International%20Conference%20on%20Data%20Engineering%20(ICDE)&rft.au=Wang,%20Sinan&rft.date=2022-05&rft.spage=3307&rft.epage=3319&rft.pages=3307-3319&rft.eissn=2375-026X&rft.coden=IEEPAD&rft_id=info:doi/10.1109/ICDE53745.2022.00312&rft.eisbn=1665408839&rft.eisbn_list=9781665408837&rft_dat=%3Cieee_CHZPO%3E9835373%3C/ieee_CHZPO%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-i203t-e8d52adc140406658ab798dc4deb8ecd7aac026ef7cf152a1ab13f4bb0c5cbab3%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_id=info:pmid/&rft_ieee_id=9835373&rfr_iscdi=true