Loading…

Emerging trends: Deep nets for poets

Deep nets have done well with early adopters, but the future will soon depend on crossing the chasm. The goal of this paper is to make deep nets more accessible to a broader audience including people with little or no programming skills, and people with little interest in training new models. A gith...

Full description

Saved in:
Bibliographic Details
Published in:Natural language engineering 2021-09, Vol.27 (5), p.631-645
Main Authors: Church, Kenneth Ward, Yuan, Xiaopeng, Guo, Sheng, Wu, Zewu, Yang, Yehua, Chen, Zeyu
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Items that cite this one
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by cdi_FETCH-LOGICAL-c360t-a1692d95bd93c0d44b6696ad15109619ea0a02438644c5c5fe4f82a45c06af813
cites cdi_FETCH-LOGICAL-c360t-a1692d95bd93c0d44b6696ad15109619ea0a02438644c5c5fe4f82a45c06af813
container_end_page 645
container_issue 5
container_start_page 631
container_title Natural language engineering
container_volume 27
creator Church, Kenneth Ward
Yuan, Xiaopeng
Guo, Sheng
Wu, Zewu
Yang, Yehua
Chen, Zeyu
description Deep nets have done well with early adopters, but the future will soon depend on crossing the chasm. The goal of this paper is to make deep nets more accessible to a broader audience including people with little or no programming skills, and people with little interest in training new models. A github is provided with simple implementations of image classification, optical character recognition, sentiment analysis, named entity recognition, question answering (QA/SQuAD), machine translation, speech to text (SST), and speech recognition (STT). The emphasis is on instant gratification. Non-programmers should be able to install these programs and use them in 15 minutes or less (per program). Programs are short (10–100 lines each) and readable by users with modest programming skills. Much of the complexity is hidden behind abstractions such as pipelines and auto classes, and pretrained models and datasets provided by hubs: PaddleHub, PaddleNLP, HuggingFaceHub, and Fairseq. Hubs have different priorities than research. Research is training models from corpora and fine-tuning them for tasks. Users are already overwhelmed with an embarrassment of riches (13k models and 1k datasets). Do they want more? We believe the broader market is more interested in inference (how to run pretrained models on novel inputs) and less interested in training (how to create even more models).
doi_str_mv 10.1017/S1351324921000231
format article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_2567804637</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><cupid>10_1017_S1351324921000231</cupid><sourcerecordid>2567804637</sourcerecordid><originalsourceid>FETCH-LOGICAL-c360t-a1692d95bd93c0d44b6696ad15109619ea0a02438644c5c5fe4f82a45c06af813</originalsourceid><addsrcrecordid>eNp1kEtLw0AUhQdRsFZ_gLuAbqP3zisZd1LrAwou1PUwmUdIsUmcSRf-e6e04EJc3QPnO-fCIeQS4QYBq9s3ZAIZ5YoiAFCGR2SGXKqyRoTjrLNd7vxTcpbSOjMcKz4j18uNj23Xt8UUfe_SXfHg_Vj0fkpFGGIxDlmdk5NgPpO_ONw5-Xhcvi-ey9Xr08viflVaJmEqDUpFnRKNU8yC47yRUknjUCAoicobMEA5qyXnVlgRPA81NVxYkCbUyObkat87xuFr69Ok18M29vmlpkJWNXDJqkzhnrJxSCn6oMfYbUz81gh6N4b-M0bOsEPGbJrYudb_Vv-f-gFFAV2G</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2567804637</pqid></control><display><type>article</type><title>Emerging trends: Deep nets for poets</title><source>Linguistics Collection</source><source>Cambridge University Press</source><source>Linguistics and Language Behavior Abstracts (LLBA)</source><source>ProQuest Social Science Premium Collection</source><creator>Church, Kenneth Ward ; Yuan, Xiaopeng ; Guo, Sheng ; Wu, Zewu ; Yang, Yehua ; Chen, Zeyu</creator><creatorcontrib>Church, Kenneth Ward ; Yuan, Xiaopeng ; Guo, Sheng ; Wu, Zewu ; Yang, Yehua ; Chen, Zeyu</creatorcontrib><description>Deep nets have done well with early adopters, but the future will soon depend on crossing the chasm. The goal of this paper is to make deep nets more accessible to a broader audience including people with little or no programming skills, and people with little interest in training new models. A github is provided with simple implementations of image classification, optical character recognition, sentiment analysis, named entity recognition, question answering (QA/SQuAD), machine translation, speech to text (SST), and speech recognition (STT). The emphasis is on instant gratification. Non-programmers should be able to install these programs and use them in 15 minutes or less (per program). Programs are short (10–100 lines each) and readable by users with modest programming skills. Much of the complexity is hidden behind abstractions such as pipelines and auto classes, and pretrained models and datasets provided by hubs: PaddleHub, PaddleNLP, HuggingFaceHub, and Fairseq. Hubs have different priorities than research. Research is training models from corpora and fine-tuning them for tasks. Users are already overwhelmed with an embarrassment of riches (13k models and 1k datasets). Do they want more? We believe the broader market is more interested in inference (how to run pretrained models on novel inputs) and less interested in training (how to create even more models).</description><identifier>ISSN: 1351-3249</identifier><identifier>EISSN: 1469-8110</identifier><identifier>DOI: 10.1017/S1351324921000231</identifier><language>eng</language><publisher>Cambridge, UK: Cambridge University Press</publisher><subject>Audiences ; Data mining ; Datasets ; Emerging Trends ; Hubs ; Ideograph recognition ; Image classification ; Inference ; Input output ; Language ; Linguistics ; Machine translation ; Object recognition ; Operating systems ; Optical character recognition ; Programmers ; Skills ; Speech ; Speech recognition ; Training ; User interface ; Voice recognition</subject><ispartof>Natural language engineering, 2021-09, Vol.27 (5), p.631-645</ispartof><rights>The Author(s), 2021. Published by Cambridge University Press</rights><rights>The Author(s), 2021. Published by Cambridge University Press. This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted re-use, distribution, and reproduction in any medium, provided the original work is properly cited. Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the associated terms available at: https://uk.sagepub.com/en-gb/eur/reusing-open-access-and-sage-choice-content</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c360t-a1692d95bd93c0d44b6696ad15109619ea0a02438644c5c5fe4f82a45c06af813</citedby><cites>FETCH-LOGICAL-c360t-a1692d95bd93c0d44b6696ad15109619ea0a02438644c5c5fe4f82a45c06af813</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://www.proquest.com/docview/2567804637/fulltextPDF?pq-origsite=primo$$EPDF$$P50$$Gproquest$$H</linktopdf><linktohtml>$$Uhttps://www.proquest.com/docview/2567804637?pq-origsite=primo$$EHTML$$P50$$Gproquest$$H</linktohtml><link.rule.ids>314,776,780,12830,21361,21373,27901,27902,31246,33588,33888,43709,43872,72703,73964,74156</link.rule.ids></links><search><creatorcontrib>Church, Kenneth Ward</creatorcontrib><creatorcontrib>Yuan, Xiaopeng</creatorcontrib><creatorcontrib>Guo, Sheng</creatorcontrib><creatorcontrib>Wu, Zewu</creatorcontrib><creatorcontrib>Yang, Yehua</creatorcontrib><creatorcontrib>Chen, Zeyu</creatorcontrib><title>Emerging trends: Deep nets for poets</title><title>Natural language engineering</title><addtitle>Nat. Lang. Eng</addtitle><description>Deep nets have done well with early adopters, but the future will soon depend on crossing the chasm. The goal of this paper is to make deep nets more accessible to a broader audience including people with little or no programming skills, and people with little interest in training new models. A github is provided with simple implementations of image classification, optical character recognition, sentiment analysis, named entity recognition, question answering (QA/SQuAD), machine translation, speech to text (SST), and speech recognition (STT). The emphasis is on instant gratification. Non-programmers should be able to install these programs and use them in 15 minutes or less (per program). Programs are short (10–100 lines each) and readable by users with modest programming skills. Much of the complexity is hidden behind abstractions such as pipelines and auto classes, and pretrained models and datasets provided by hubs: PaddleHub, PaddleNLP, HuggingFaceHub, and Fairseq. Hubs have different priorities than research. Research is training models from corpora and fine-tuning them for tasks. Users are already overwhelmed with an embarrassment of riches (13k models and 1k datasets). Do they want more? We believe the broader market is more interested in inference (how to run pretrained models on novel inputs) and less interested in training (how to create even more models).</description><subject>Audiences</subject><subject>Data mining</subject><subject>Datasets</subject><subject>Emerging Trends</subject><subject>Hubs</subject><subject>Ideograph recognition</subject><subject>Image classification</subject><subject>Inference</subject><subject>Input output</subject><subject>Language</subject><subject>Linguistics</subject><subject>Machine translation</subject><subject>Object recognition</subject><subject>Operating systems</subject><subject>Optical character recognition</subject><subject>Programmers</subject><subject>Skills</subject><subject>Speech</subject><subject>Speech recognition</subject><subject>Training</subject><subject>User interface</subject><subject>Voice recognition</subject><issn>1351-3249</issn><issn>1469-8110</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2021</creationdate><recordtype>article</recordtype><sourceid>7T9</sourceid><sourceid>ALSLI</sourceid><sourceid>CPGLG</sourceid><sourceid>CRLPW</sourceid><recordid>eNp1kEtLw0AUhQdRsFZ_gLuAbqP3zisZd1LrAwou1PUwmUdIsUmcSRf-e6e04EJc3QPnO-fCIeQS4QYBq9s3ZAIZ5YoiAFCGR2SGXKqyRoTjrLNd7vxTcpbSOjMcKz4j18uNj23Xt8UUfe_SXfHg_Vj0fkpFGGIxDlmdk5NgPpO_ONw5-Xhcvi-ey9Xr08viflVaJmEqDUpFnRKNU8yC47yRUknjUCAoicobMEA5qyXnVlgRPA81NVxYkCbUyObkat87xuFr69Ok18M29vmlpkJWNXDJqkzhnrJxSCn6oMfYbUz81gh6N4b-M0bOsEPGbJrYudb_Vv-f-gFFAV2G</recordid><startdate>202109</startdate><enddate>202109</enddate><creator>Church, Kenneth Ward</creator><creator>Yuan, Xiaopeng</creator><creator>Guo, Sheng</creator><creator>Wu, Zewu</creator><creator>Yang, Yehua</creator><creator>Chen, Zeyu</creator><general>Cambridge University Press</general><scope>IKXGN</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>3V.</scope><scope>7T9</scope><scope>7XB</scope><scope>88G</scope><scope>8AL</scope><scope>8FE</scope><scope>8FG</scope><scope>8FI</scope><scope>8FJ</scope><scope>8FK</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AEUYN</scope><scope>AFKRA</scope><scope>ALSLI</scope><scope>ARAPS</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>CPGLG</scope><scope>CRLPW</scope><scope>DWQXO</scope><scope>FYUFA</scope><scope>GHDGH</scope><scope>GNUQQ</scope><scope>HCIFZ</scope><scope>JQ2</scope><scope>K7-</scope><scope>L6V</scope><scope>M0N</scope><scope>M2M</scope><scope>M7S</scope><scope>P5Z</scope><scope>P62</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PSYQQ</scope><scope>PTHSS</scope><scope>Q9U</scope></search><sort><creationdate>202109</creationdate><title>Emerging trends: Deep nets for poets</title><author>Church, Kenneth Ward ; Yuan, Xiaopeng ; Guo, Sheng ; Wu, Zewu ; Yang, Yehua ; Chen, Zeyu</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c360t-a1692d95bd93c0d44b6696ad15109619ea0a02438644c5c5fe4f82a45c06af813</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2021</creationdate><topic>Audiences</topic><topic>Data mining</topic><topic>Datasets</topic><topic>Emerging Trends</topic><topic>Hubs</topic><topic>Ideograph recognition</topic><topic>Image classification</topic><topic>Inference</topic><topic>Input output</topic><topic>Language</topic><topic>Linguistics</topic><topic>Machine translation</topic><topic>Object recognition</topic><topic>Operating systems</topic><topic>Optical character recognition</topic><topic>Programmers</topic><topic>Skills</topic><topic>Speech</topic><topic>Speech recognition</topic><topic>Training</topic><topic>User interface</topic><topic>Voice recognition</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Church, Kenneth Ward</creatorcontrib><creatorcontrib>Yuan, Xiaopeng</creatorcontrib><creatorcontrib>Guo, Sheng</creatorcontrib><creatorcontrib>Wu, Zewu</creatorcontrib><creatorcontrib>Yang, Yehua</creatorcontrib><creatorcontrib>Chen, Zeyu</creatorcontrib><collection>Cambridge University Press:Open Access Journals</collection><collection>CrossRef</collection><collection>ProQuest Central (Corporate)</collection><collection>Linguistics and Language Behavior Abstracts (LLBA)</collection><collection>ProQuest Central (purchase pre-March 2016)</collection><collection>Psychology Database (Alumni)</collection><collection>Computing Database (Alumni Edition)</collection><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Hospital Premium Collection</collection><collection>Hospital Premium Collection (Alumni Edition)</collection><collection>ProQuest Central (Alumni) (purchase pre-March 2016)</collection><collection>Materials Science &amp; Engineering Collection</collection><collection>ProQuest Central (Alumni)</collection><collection>ProQuest One Sustainability</collection><collection>ProQuest Central UK/Ireland</collection><collection>ProQuest Social Science Premium Collection</collection><collection>Advanced Technologies &amp; Aerospace Database‎ (1962 - current)</collection><collection>ProQuest Central Essentials</collection><collection>AUTh Library subscriptions: ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>Linguistics Collection</collection><collection>Linguistics Database</collection><collection>ProQuest Central</collection><collection>Health Research Premium Collection</collection><collection>Health Research Premium Collection (Alumni)</collection><collection>ProQuest Central Student</collection><collection>SciTech Premium Collection (Proquest) (PQ_SDU_P3)</collection><collection>ProQuest Computer Science Collection</collection><collection>Computer Science Database</collection><collection>ProQuest Engineering Collection</collection><collection>Computing Database</collection><collection>ProQuest Psychology</collection><collection>Engineering Database</collection><collection>ProQuest advanced technologies &amp; aerospace journals</collection><collection>ProQuest Advanced Technologies &amp; Aerospace Collection</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest One Psychology</collection><collection>Engineering collection</collection><collection>ProQuest Central Basic</collection><jtitle>Natural language engineering</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Church, Kenneth Ward</au><au>Yuan, Xiaopeng</au><au>Guo, Sheng</au><au>Wu, Zewu</au><au>Yang, Yehua</au><au>Chen, Zeyu</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Emerging trends: Deep nets for poets</atitle><jtitle>Natural language engineering</jtitle><addtitle>Nat. Lang. Eng</addtitle><date>2021-09</date><risdate>2021</risdate><volume>27</volume><issue>5</issue><spage>631</spage><epage>645</epage><pages>631-645</pages><issn>1351-3249</issn><eissn>1469-8110</eissn><abstract>Deep nets have done well with early adopters, but the future will soon depend on crossing the chasm. The goal of this paper is to make deep nets more accessible to a broader audience including people with little or no programming skills, and people with little interest in training new models. A github is provided with simple implementations of image classification, optical character recognition, sentiment analysis, named entity recognition, question answering (QA/SQuAD), machine translation, speech to text (SST), and speech recognition (STT). The emphasis is on instant gratification. Non-programmers should be able to install these programs and use them in 15 minutes or less (per program). Programs are short (10–100 lines each) and readable by users with modest programming skills. Much of the complexity is hidden behind abstractions such as pipelines and auto classes, and pretrained models and datasets provided by hubs: PaddleHub, PaddleNLP, HuggingFaceHub, and Fairseq. Hubs have different priorities than research. Research is training models from corpora and fine-tuning them for tasks. Users are already overwhelmed with an embarrassment of riches (13k models and 1k datasets). Do they want more? We believe the broader market is more interested in inference (how to run pretrained models on novel inputs) and less interested in training (how to create even more models).</abstract><cop>Cambridge, UK</cop><pub>Cambridge University Press</pub><doi>10.1017/S1351324921000231</doi><tpages>15</tpages><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 1351-3249
ispartof Natural language engineering, 2021-09, Vol.27 (5), p.631-645
issn 1351-3249
1469-8110
language eng
recordid cdi_proquest_journals_2567804637
source Linguistics Collection; Cambridge University Press; Linguistics and Language Behavior Abstracts (LLBA); ProQuest Social Science Premium Collection
subjects Audiences
Data mining
Datasets
Emerging Trends
Hubs
Ideograph recognition
Image classification
Inference
Input output
Language
Linguistics
Machine translation
Object recognition
Operating systems
Optical character recognition
Programmers
Skills
Speech
Speech recognition
Training
User interface
Voice recognition
title Emerging trends: Deep nets for poets
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-12T12%3A07%3A25IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Emerging%20trends:%20Deep%20nets%20for%20poets&rft.jtitle=Natural%20language%20engineering&rft.au=Church,%20Kenneth%20Ward&rft.date=2021-09&rft.volume=27&rft.issue=5&rft.spage=631&rft.epage=645&rft.pages=631-645&rft.issn=1351-3249&rft.eissn=1469-8110&rft_id=info:doi/10.1017/S1351324921000231&rft_dat=%3Cproquest_cross%3E2567804637%3C/proquest_cross%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c360t-a1692d95bd93c0d44b6696ad15109619ea0a02438644c5c5fe4f82a45c06af813%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2567804637&rft_id=info:pmid/&rft_cupid=10_1017_S1351324921000231&rfr_iscdi=true