Loading…

Self-Supervised Time Series Representation Learning via Cross Reconstruction Transformer

Since labeled samples are typically scarce in real-world scenarios, self-supervised representation learning in time series is critical. Existing approaches mainly employ the contrastive learning framework, which automatically learns to understand similar and dissimilar data pairs. However, they are...

Full description

Saved in:
Bibliographic Details
Published in:IEEE transaction on neural networks and learning systems 2024-11, Vol.35 (11), p.16129-16138
Main Authors: Zhang, Wenrui, Yang, Ling, Geng, Shijia, Hong, Shenda
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Items that cite this one
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by cdi_FETCH-LOGICAL-c324t-7c459486dd7b6306f32885941b9465ced87d568faa7db685eb0a8f01fd3968ab3
cites cdi_FETCH-LOGICAL-c324t-7c459486dd7b6306f32885941b9465ced87d568faa7db685eb0a8f01fd3968ab3
container_end_page 16138
container_issue 11
container_start_page 16129
container_title IEEE transaction on neural networks and learning systems
container_volume 35
creator Zhang, Wenrui
Yang, Ling
Geng, Shijia
Hong, Shenda
description Since labeled samples are typically scarce in real-world scenarios, self-supervised representation learning in time series is critical. Existing approaches mainly employ the contrastive learning framework, which automatically learns to understand similar and dissimilar data pairs. However, they are constrained by the request for cumbersome sampling policies and prior knowledge of constructing pairs. Also, few works have focused on effectively modeling temporal-spectral correlations to improve the capacity of representations. In this article, we propose the cross reconstruction transformer (CRT) to solve the aforementioned issues. CRT achieves time series representation learning through a cross-domain dropping-reconstruction task. Specifically, we obtain the frequency domain of the time series via the fast Fourier transform (FFT) and randomly drop certain patches in both time and frequency domains. Dropping is employed to maximally preserve the global context while masking leads to the distribution shift. Then a Transformer architecture is utilized to adequately discover the cross-domain correlations between temporal and spectral information through reconstructing data in both domains, which is called Dropped Temporal-Spectral Modeling. To discriminate the representations in global latent space, we propose instance discrimination constraint (IDC) to reduce the mutual information between different time series samples and sharpen the decision boundaries. Additionally, a specified curriculum learning (CL) strategy is employed to improve the robustness during the pretraining phase, which progressively increases the dropping ratio in the training process. We conduct extensive experiments to evaluate the effectiveness of the proposed method on multiple real-world datasets. Results show that CRT consistently achieves the best performance over existing methods by 2%-9%. The code is publicly available at https://github.com/BobZwr/Cross-Reconstruction-Transformer .
doi_str_mv 10.1109/TNNLS.2023.3292066
format article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_crossref_primary_10_1109_TNNLS_2023_3292066</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>10190201</ieee_id><sourcerecordid>2841027747</sourcerecordid><originalsourceid>FETCH-LOGICAL-c324t-7c459486dd7b6306f32885941b9465ced87d568faa7db685eb0a8f01fd3968ab3</originalsourceid><addsrcrecordid>eNpNkFtLwzAYhoMobsz9ARHppTedObRpeinDE5QJtsLuQtp8lUhPJu3Af2-7zWFuEsLzvnzfg9A1wStCcHyfbTZJuqKYshWjMcWcn6E5JZz6lAlxfnpH2xlaOveFx8NxyIP4Es1YFEQCB3SOtilUpZ8OHdidcaC9zNTgpWANOO8dOgsOml71pm28BJRtTPPp7Yzy1rZ1E1G0jevtUOyJzKrGla2twV6hi1JVDpbHe4E-nh6z9YufvD2_rh8Sv2A06P2oCMI4EFzrKOcM85JRIcYfkscBDwvQItIhF6VSkc65CCHHSpSYlJrFXKicLdDdobez7fcArpe1cQVUlWqgHZykIiCYRuPCI0oPaDHNbqGUnTW1sj-SYDlJlXupcpIqj1LH0O2xf8hr0KfIn8IRuDkABgD-NZIYU0zYL9WxfFA</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2841027747</pqid></control><display><type>article</type><title>Self-Supervised Time Series Representation Learning via Cross Reconstruction Transformer</title><source>IEEE Electronic Library (IEL) Journals</source><creator>Zhang, Wenrui ; Yang, Ling ; Geng, Shijia ; Hong, Shenda</creator><creatorcontrib>Zhang, Wenrui ; Yang, Ling ; Geng, Shijia ; Hong, Shenda</creatorcontrib><description>Since labeled samples are typically scarce in real-world scenarios, self-supervised representation learning in time series is critical. Existing approaches mainly employ the contrastive learning framework, which automatically learns to understand similar and dissimilar data pairs. However, they are constrained by the request for cumbersome sampling policies and prior knowledge of constructing pairs. Also, few works have focused on effectively modeling temporal-spectral correlations to improve the capacity of representations. In this article, we propose the cross reconstruction transformer (CRT) to solve the aforementioned issues. CRT achieves time series representation learning through a cross-domain dropping-reconstruction task. Specifically, we obtain the frequency domain of the time series via the fast Fourier transform (FFT) and randomly drop certain patches in both time and frequency domains. Dropping is employed to maximally preserve the global context while masking leads to the distribution shift. Then a Transformer architecture is utilized to adequately discover the cross-domain correlations between temporal and spectral information through reconstructing data in both domains, which is called Dropped Temporal-Spectral Modeling. To discriminate the representations in global latent space, we propose instance discrimination constraint (IDC) to reduce the mutual information between different time series samples and sharpen the decision boundaries. Additionally, a specified curriculum learning (CL) strategy is employed to improve the robustness during the pretraining phase, which progressively increases the dropping ratio in the training process. We conduct extensive experiments to evaluate the effectiveness of the proposed method on multiple real-world datasets. Results show that CRT consistently achieves the best performance over existing methods by 2%-9%. The code is publicly available at https://github.com/BobZwr/Cross-Reconstruction-Transformer .</description><identifier>ISSN: 2162-237X</identifier><identifier>ISSN: 2162-2388</identifier><identifier>EISSN: 2162-2388</identifier><identifier>DOI: 10.1109/TNNLS.2023.3292066</identifier><identifier>PMID: 37478042</identifier><identifier>CODEN: ITNNAL</identifier><language>eng</language><publisher>United States: IEEE</publisher><subject>Cathode ray tubes ; Cross domain ; Image reconstruction ; Representation learning ; Self-supervised learning ; Task analysis ; time series ; Time series analysis ; transformer ; Transformers</subject><ispartof>IEEE transaction on neural networks and learning systems, 2024-11, Vol.35 (11), p.16129-16138</ispartof><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c324t-7c459486dd7b6306f32885941b9465ced87d568faa7db685eb0a8f01fd3968ab3</citedby><cites>FETCH-LOGICAL-c324t-7c459486dd7b6306f32885941b9465ced87d568faa7db685eb0a8f01fd3968ab3</cites><orcidid>0000-0003-1905-8053 ; 0000-0001-6972-7038 ; 0000-0002-9628-8808 ; 0000-0001-7521-5127</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/10190201$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>314,776,780,27901,27902,54771</link.rule.ids><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/37478042$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Zhang, Wenrui</creatorcontrib><creatorcontrib>Yang, Ling</creatorcontrib><creatorcontrib>Geng, Shijia</creatorcontrib><creatorcontrib>Hong, Shenda</creatorcontrib><title>Self-Supervised Time Series Representation Learning via Cross Reconstruction Transformer</title><title>IEEE transaction on neural networks and learning systems</title><addtitle>TNNLS</addtitle><addtitle>IEEE Trans Neural Netw Learn Syst</addtitle><description>Since labeled samples are typically scarce in real-world scenarios, self-supervised representation learning in time series is critical. Existing approaches mainly employ the contrastive learning framework, which automatically learns to understand similar and dissimilar data pairs. However, they are constrained by the request for cumbersome sampling policies and prior knowledge of constructing pairs. Also, few works have focused on effectively modeling temporal-spectral correlations to improve the capacity of representations. In this article, we propose the cross reconstruction transformer (CRT) to solve the aforementioned issues. CRT achieves time series representation learning through a cross-domain dropping-reconstruction task. Specifically, we obtain the frequency domain of the time series via the fast Fourier transform (FFT) and randomly drop certain patches in both time and frequency domains. Dropping is employed to maximally preserve the global context while masking leads to the distribution shift. Then a Transformer architecture is utilized to adequately discover the cross-domain correlations between temporal and spectral information through reconstructing data in both domains, which is called Dropped Temporal-Spectral Modeling. To discriminate the representations in global latent space, we propose instance discrimination constraint (IDC) to reduce the mutual information between different time series samples and sharpen the decision boundaries. Additionally, a specified curriculum learning (CL) strategy is employed to improve the robustness during the pretraining phase, which progressively increases the dropping ratio in the training process. We conduct extensive experiments to evaluate the effectiveness of the proposed method on multiple real-world datasets. Results show that CRT consistently achieves the best performance over existing methods by 2%-9%. The code is publicly available at https://github.com/BobZwr/Cross-Reconstruction-Transformer .</description><subject>Cathode ray tubes</subject><subject>Cross domain</subject><subject>Image reconstruction</subject><subject>Representation learning</subject><subject>Self-supervised learning</subject><subject>Task analysis</subject><subject>time series</subject><subject>Time series analysis</subject><subject>transformer</subject><subject>Transformers</subject><issn>2162-237X</issn><issn>2162-2388</issn><issn>2162-2388</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><recordid>eNpNkFtLwzAYhoMobsz9ARHppTedObRpeinDE5QJtsLuQtp8lUhPJu3Af2-7zWFuEsLzvnzfg9A1wStCcHyfbTZJuqKYshWjMcWcn6E5JZz6lAlxfnpH2xlaOveFx8NxyIP4Es1YFEQCB3SOtilUpZ8OHdidcaC9zNTgpWANOO8dOgsOml71pm28BJRtTPPp7Yzy1rZ1E1G0jevtUOyJzKrGla2twV6hi1JVDpbHe4E-nh6z9YufvD2_rh8Sv2A06P2oCMI4EFzrKOcM85JRIcYfkscBDwvQItIhF6VSkc65CCHHSpSYlJrFXKicLdDdobez7fcArpe1cQVUlWqgHZykIiCYRuPCI0oPaDHNbqGUnTW1sj-SYDlJlXupcpIqj1LH0O2xf8hr0KfIn8IRuDkABgD-NZIYU0zYL9WxfFA</recordid><startdate>20241101</startdate><enddate>20241101</enddate><creator>Zhang, Wenrui</creator><creator>Yang, Ling</creator><creator>Geng, Shijia</creator><creator>Hong, Shenda</creator><general>IEEE</general><scope>97E</scope><scope>RIA</scope><scope>RIE</scope><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7X8</scope><orcidid>https://orcid.org/0000-0003-1905-8053</orcidid><orcidid>https://orcid.org/0000-0001-6972-7038</orcidid><orcidid>https://orcid.org/0000-0002-9628-8808</orcidid><orcidid>https://orcid.org/0000-0001-7521-5127</orcidid></search><sort><creationdate>20241101</creationdate><title>Self-Supervised Time Series Representation Learning via Cross Reconstruction Transformer</title><author>Zhang, Wenrui ; Yang, Ling ; Geng, Shijia ; Hong, Shenda</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c324t-7c459486dd7b6306f32885941b9465ced87d568faa7db685eb0a8f01fd3968ab3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Cathode ray tubes</topic><topic>Cross domain</topic><topic>Image reconstruction</topic><topic>Representation learning</topic><topic>Self-supervised learning</topic><topic>Task analysis</topic><topic>time series</topic><topic>Time series analysis</topic><topic>transformer</topic><topic>Transformers</topic><toplevel>online_resources</toplevel><creatorcontrib>Zhang, Wenrui</creatorcontrib><creatorcontrib>Yang, Ling</creatorcontrib><creatorcontrib>Geng, Shijia</creatorcontrib><creatorcontrib>Hong, Shenda</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005–Present</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998–Present</collection><collection>IEEE Xplore</collection><collection>PubMed</collection><collection>CrossRef</collection><collection>MEDLINE - Academic</collection><jtitle>IEEE transaction on neural networks and learning systems</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Zhang, Wenrui</au><au>Yang, Ling</au><au>Geng, Shijia</au><au>Hong, Shenda</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Self-Supervised Time Series Representation Learning via Cross Reconstruction Transformer</atitle><jtitle>IEEE transaction on neural networks and learning systems</jtitle><stitle>TNNLS</stitle><addtitle>IEEE Trans Neural Netw Learn Syst</addtitle><date>2024-11-01</date><risdate>2024</risdate><volume>35</volume><issue>11</issue><spage>16129</spage><epage>16138</epage><pages>16129-16138</pages><issn>2162-237X</issn><issn>2162-2388</issn><eissn>2162-2388</eissn><coden>ITNNAL</coden><abstract>Since labeled samples are typically scarce in real-world scenarios, self-supervised representation learning in time series is critical. Existing approaches mainly employ the contrastive learning framework, which automatically learns to understand similar and dissimilar data pairs. However, they are constrained by the request for cumbersome sampling policies and prior knowledge of constructing pairs. Also, few works have focused on effectively modeling temporal-spectral correlations to improve the capacity of representations. In this article, we propose the cross reconstruction transformer (CRT) to solve the aforementioned issues. CRT achieves time series representation learning through a cross-domain dropping-reconstruction task. Specifically, we obtain the frequency domain of the time series via the fast Fourier transform (FFT) and randomly drop certain patches in both time and frequency domains. Dropping is employed to maximally preserve the global context while masking leads to the distribution shift. Then a Transformer architecture is utilized to adequately discover the cross-domain correlations between temporal and spectral information through reconstructing data in both domains, which is called Dropped Temporal-Spectral Modeling. To discriminate the representations in global latent space, we propose instance discrimination constraint (IDC) to reduce the mutual information between different time series samples and sharpen the decision boundaries. Additionally, a specified curriculum learning (CL) strategy is employed to improve the robustness during the pretraining phase, which progressively increases the dropping ratio in the training process. We conduct extensive experiments to evaluate the effectiveness of the proposed method on multiple real-world datasets. Results show that CRT consistently achieves the best performance over existing methods by 2%-9%. The code is publicly available at https://github.com/BobZwr/Cross-Reconstruction-Transformer .</abstract><cop>United States</cop><pub>IEEE</pub><pmid>37478042</pmid><doi>10.1109/TNNLS.2023.3292066</doi><tpages>10</tpages><orcidid>https://orcid.org/0000-0003-1905-8053</orcidid><orcidid>https://orcid.org/0000-0001-6972-7038</orcidid><orcidid>https://orcid.org/0000-0002-9628-8808</orcidid><orcidid>https://orcid.org/0000-0001-7521-5127</orcidid></addata></record>
fulltext fulltext
identifier ISSN: 2162-237X
ispartof IEEE transaction on neural networks and learning systems, 2024-11, Vol.35 (11), p.16129-16138
issn 2162-237X
2162-2388
2162-2388
language eng
recordid cdi_crossref_primary_10_1109_TNNLS_2023_3292066
source IEEE Electronic Library (IEL) Journals
subjects Cathode ray tubes
Cross domain
Image reconstruction
Representation learning
Self-supervised learning
Task analysis
time series
Time series analysis
transformer
Transformers
title Self-Supervised Time Series Representation Learning via Cross Reconstruction Transformer
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-01T01%3A04%3A40IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Self-Supervised%20Time%20Series%20Representation%20Learning%20via%20Cross%20Reconstruction%20Transformer&rft.jtitle=IEEE%20transaction%20on%20neural%20networks%20and%20learning%20systems&rft.au=Zhang,%20Wenrui&rft.date=2024-11-01&rft.volume=35&rft.issue=11&rft.spage=16129&rft.epage=16138&rft.pages=16129-16138&rft.issn=2162-237X&rft.eissn=2162-2388&rft.coden=ITNNAL&rft_id=info:doi/10.1109/TNNLS.2023.3292066&rft_dat=%3Cproquest_cross%3E2841027747%3C/proquest_cross%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c324t-7c459486dd7b6306f32885941b9465ced87d568faa7db685eb0a8f01fd3968ab3%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2841027747&rft_id=info:pmid/37478042&rft_ieee_id=10190201&rfr_iscdi=true