Loading…

Application of Transformers in Cheminformatics

By accelerating time-consuming processes with high efficiency, computing has become an essential part of many modern chemical pipelines. Machine learning is a class of computing methods that can discover patterns within chemical data and utilize this knowledge for a wide variety of downstream tasks,...

Full description

Saved in:
Bibliographic Details
Published in:Journal of chemical information and modeling 2024-06, Vol.64 (11), p.4392-4409
Main Authors: Luong, Kha-Dinh, Singh, Ambuj
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by
cites cdi_FETCH-LOGICAL-a415t-373b3bd04561623ef8aef96a5f2d3ec6ae673040fc72e52456568163917ced83
container_end_page 4409
container_issue 11
container_start_page 4392
container_title Journal of chemical information and modeling
container_volume 64
creator Luong, Kha-Dinh
Singh, Ambuj
description By accelerating time-consuming processes with high efficiency, computing has become an essential part of many modern chemical pipelines. Machine learning is a class of computing methods that can discover patterns within chemical data and utilize this knowledge for a wide variety of downstream tasks, such as property prediction or substance generation. The complex and diverse chemical space requires complex machine learning architectures with great learning power. Recently, learning models based on transformer architectures have revolutionized multiple domains of machine learning, including natural language processing and computer vision. Naturally, there have been ongoing endeavors in adopting these techniques to the chemical domain, resulting in a surge of publications within a short period. The diversity of chemical structures, use cases, and learning models necessitate a comprehensive summarization of existing works. In this paper, we review recent innovations in adapting transformers to solve learning problems in chemistry. Because chemical data is diverse and complex, we structure our discussion based on chemical representations. Specifically, we highlight the strengths and weaknesses of each representation, the current progress of adapting transformer architectures, and future directions.
doi_str_mv 10.1021/acs.jcim.3c02070
format article
fullrecord <record><control><sourceid>proquest_pubme</sourceid><recordid>TN_cdi_pubmedcentral_primary_oai_pubmedcentral_nih_gov_11167597</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>3066505660</sourcerecordid><originalsourceid>FETCH-LOGICAL-a415t-373b3bd04561623ef8aef96a5f2d3ec6ae673040fc72e52456568163917ced83</originalsourceid><addsrcrecordid>eNp1kc1LxDAQxYMorq7ePUnBiwdbk6aZtidZFr9gwcsevIVsmrhZ2qYmXcH_3tT9QAVPCZnfezOTh9AFwQnBKbkV0icraZqESpziHB-gE8KyMi4Bvx7u7qyEETr1foUxpSWkx2hEi4KwNIMTlEy6rjZS9Ma2kdXR3InWa-sa5Xxk2mi6VI1ph4eASH-GjrSovTrfnmM0f7ifT5_i2cvj83Qyi0VGWB_TnC7oosIZAwIpVboQSpcgmE4rqiQIBTnFGdYyT1UYhAGDggAtSS5VVdAxutvYdutFoyqp2t6JmnfONMJ9cisM_11pzZK_2Q9OCIGclXlwuN46OPu-Vr7njfFS1bVolV17TjHQjOUFZAG9-oOu7Nq1Yb2BAoYZAA4U3lDSWe-d0vtpCOZDGDyEwYcw-DaMILn8ucVesPv9ANxsgG_prum_fl9x_pUl</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>3066505660</pqid></control><display><type>article</type><title>Application of Transformers in Cheminformatics</title><source>American Chemical Society:Jisc Collections:American Chemical Society Read &amp; Publish Agreement 2022-2024 (Reading list)</source><creator>Luong, Kha-Dinh ; Singh, Ambuj</creator><creatorcontrib>Luong, Kha-Dinh ; Singh, Ambuj</creatorcontrib><description>By accelerating time-consuming processes with high efficiency, computing has become an essential part of many modern chemical pipelines. Machine learning is a class of computing methods that can discover patterns within chemical data and utilize this knowledge for a wide variety of downstream tasks, such as property prediction or substance generation. The complex and diverse chemical space requires complex machine learning architectures with great learning power. Recently, learning models based on transformer architectures have revolutionized multiple domains of machine learning, including natural language processing and computer vision. Naturally, there have been ongoing endeavors in adopting these techniques to the chemical domain, resulting in a surge of publications within a short period. The diversity of chemical structures, use cases, and learning models necessitate a comprehensive summarization of existing works. In this paper, we review recent innovations in adapting transformers to solve learning problems in chemistry. Because chemical data is diverse and complex, we structure our discussion based on chemical representations. Specifically, we highlight the strengths and weaknesses of each representation, the current progress of adapting transformer architectures, and future directions.</description><identifier>ISSN: 1549-9596</identifier><identifier>ISSN: 1549-960X</identifier><identifier>EISSN: 1549-960X</identifier><identifier>DOI: 10.1021/acs.jcim.3c02070</identifier><identifier>PMID: 38815246</identifier><language>eng</language><publisher>United States: American Chemical Society</publisher><subject>Computation ; Computer vision ; Machine learning ; Natural language processing ; Representations ; Review ; Task complexity</subject><ispartof>Journal of chemical information and modeling, 2024-06, Vol.64 (11), p.4392-4409</ispartof><rights>2024 The Authors. Published by American Chemical Society</rights><rights>Copyright American Chemical Society Jun 10, 2024</rights><rights>2024 The Authors. Published by American Chemical Society 2024 The Authors</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><cites>FETCH-LOGICAL-a415t-373b3bd04561623ef8aef96a5f2d3ec6ae673040fc72e52456568163917ced83</cites><orcidid>0009-0003-6919-4528</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>230,314,776,780,881,27901,27902</link.rule.ids><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/38815246$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Luong, Kha-Dinh</creatorcontrib><creatorcontrib>Singh, Ambuj</creatorcontrib><title>Application of Transformers in Cheminformatics</title><title>Journal of chemical information and modeling</title><addtitle>J. Chem. Inf. Model</addtitle><description>By accelerating time-consuming processes with high efficiency, computing has become an essential part of many modern chemical pipelines. Machine learning is a class of computing methods that can discover patterns within chemical data and utilize this knowledge for a wide variety of downstream tasks, such as property prediction or substance generation. The complex and diverse chemical space requires complex machine learning architectures with great learning power. Recently, learning models based on transformer architectures have revolutionized multiple domains of machine learning, including natural language processing and computer vision. Naturally, there have been ongoing endeavors in adopting these techniques to the chemical domain, resulting in a surge of publications within a short period. The diversity of chemical structures, use cases, and learning models necessitate a comprehensive summarization of existing works. In this paper, we review recent innovations in adapting transformers to solve learning problems in chemistry. Because chemical data is diverse and complex, we structure our discussion based on chemical representations. Specifically, we highlight the strengths and weaknesses of each representation, the current progress of adapting transformer architectures, and future directions.</description><subject>Computation</subject><subject>Computer vision</subject><subject>Machine learning</subject><subject>Natural language processing</subject><subject>Representations</subject><subject>Review</subject><subject>Task complexity</subject><issn>1549-9596</issn><issn>1549-960X</issn><issn>1549-960X</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><recordid>eNp1kc1LxDAQxYMorq7ePUnBiwdbk6aZtidZFr9gwcsevIVsmrhZ2qYmXcH_3tT9QAVPCZnfezOTh9AFwQnBKbkV0icraZqESpziHB-gE8KyMi4Bvx7u7qyEETr1foUxpSWkx2hEi4KwNIMTlEy6rjZS9Ma2kdXR3InWa-sa5Xxk2mi6VI1ph4eASH-GjrSovTrfnmM0f7ifT5_i2cvj83Qyi0VGWB_TnC7oosIZAwIpVboQSpcgmE4rqiQIBTnFGdYyT1UYhAGDggAtSS5VVdAxutvYdutFoyqp2t6JmnfONMJ9cisM_11pzZK_2Q9OCIGclXlwuN46OPu-Vr7njfFS1bVolV17TjHQjOUFZAG9-oOu7Nq1Yb2BAoYZAA4U3lDSWe-d0vtpCOZDGDyEwYcw-DaMILn8ucVesPv9ANxsgG_prum_fl9x_pUl</recordid><startdate>20240610</startdate><enddate>20240610</enddate><creator>Luong, Kha-Dinh</creator><creator>Singh, Ambuj</creator><general>American Chemical Society</general><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>7SR</scope><scope>7U5</scope><scope>8BQ</scope><scope>8FD</scope><scope>JG9</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>7X8</scope><scope>5PM</scope><orcidid>https://orcid.org/0009-0003-6919-4528</orcidid></search><sort><creationdate>20240610</creationdate><title>Application of Transformers in Cheminformatics</title><author>Luong, Kha-Dinh ; Singh, Ambuj</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a415t-373b3bd04561623ef8aef96a5f2d3ec6ae673040fc72e52456568163917ced83</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Computation</topic><topic>Computer vision</topic><topic>Machine learning</topic><topic>Natural language processing</topic><topic>Representations</topic><topic>Review</topic><topic>Task complexity</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Luong, Kha-Dinh</creatorcontrib><creatorcontrib>Singh, Ambuj</creatorcontrib><collection>PubMed</collection><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Engineered Materials Abstracts</collection><collection>Solid State and Superconductivity Abstracts</collection><collection>METADEX</collection><collection>Technology Research Database</collection><collection>Materials Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>MEDLINE - Academic</collection><collection>PubMed Central (Full Participant titles)</collection><jtitle>Journal of chemical information and modeling</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Luong, Kha-Dinh</au><au>Singh, Ambuj</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Application of Transformers in Cheminformatics</atitle><jtitle>Journal of chemical information and modeling</jtitle><addtitle>J. Chem. Inf. Model</addtitle><date>2024-06-10</date><risdate>2024</risdate><volume>64</volume><issue>11</issue><spage>4392</spage><epage>4409</epage><pages>4392-4409</pages><issn>1549-9596</issn><issn>1549-960X</issn><eissn>1549-960X</eissn><abstract>By accelerating time-consuming processes with high efficiency, computing has become an essential part of many modern chemical pipelines. Machine learning is a class of computing methods that can discover patterns within chemical data and utilize this knowledge for a wide variety of downstream tasks, such as property prediction or substance generation. The complex and diverse chemical space requires complex machine learning architectures with great learning power. Recently, learning models based on transformer architectures have revolutionized multiple domains of machine learning, including natural language processing and computer vision. Naturally, there have been ongoing endeavors in adopting these techniques to the chemical domain, resulting in a surge of publications within a short period. The diversity of chemical structures, use cases, and learning models necessitate a comprehensive summarization of existing works. In this paper, we review recent innovations in adapting transformers to solve learning problems in chemistry. Because chemical data is diverse and complex, we structure our discussion based on chemical representations. Specifically, we highlight the strengths and weaknesses of each representation, the current progress of adapting transformer architectures, and future directions.</abstract><cop>United States</cop><pub>American Chemical Society</pub><pmid>38815246</pmid><doi>10.1021/acs.jcim.3c02070</doi><tpages>18</tpages><orcidid>https://orcid.org/0009-0003-6919-4528</orcidid><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 1549-9596
ispartof Journal of chemical information and modeling, 2024-06, Vol.64 (11), p.4392-4409
issn 1549-9596
1549-960X
1549-960X
language eng
recordid cdi_pubmedcentral_primary_oai_pubmedcentral_nih_gov_11167597
source American Chemical Society:Jisc Collections:American Chemical Society Read & Publish Agreement 2022-2024 (Reading list)
subjects Computation
Computer vision
Machine learning
Natural language processing
Representations
Review
Task complexity
title Application of Transformers in Cheminformatics
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-10T13%3A19%3A42IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_pubme&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Application%20of%20Transformers%20in%20Cheminformatics&rft.jtitle=Journal%20of%20chemical%20information%20and%20modeling&rft.au=Luong,%20Kha-Dinh&rft.date=2024-06-10&rft.volume=64&rft.issue=11&rft.spage=4392&rft.epage=4409&rft.pages=4392-4409&rft.issn=1549-9596&rft.eissn=1549-960X&rft_id=info:doi/10.1021/acs.jcim.3c02070&rft_dat=%3Cproquest_pubme%3E3066505660%3C/proquest_pubme%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-a415t-373b3bd04561623ef8aef96a5f2d3ec6ae673040fc72e52456568163917ced83%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=3066505660&rft_id=info:pmid/38815246&rfr_iscdi=true