Loading…
An evaluation of recent neural sequence tagging models in Turkish named entity recognition
•Performances of LSTM and transformer-based networks are studied for Turkish NER.•A new transformer-based network with a CRF layer on top is introduced.•The state-of-the-art Turkish NER results are obtained with the proposed network Named entity recognition (NER) is an extensively studied task that...
Saved in:
Published in: | Expert systems with applications 2021-11, Vol.182, p.115049, Article 115049 |
---|---|
Main Authors: | , , , |
Format: | Article |
Language: | English |
Subjects: | |
Citations: | Items that this one cites Items that cite this one |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
cited_by | cdi_FETCH-LOGICAL-c372t-b0693cc10f586ca08dfb6b81f6ba59c249199f572c4d93465854b9c134fd619b3 |
---|---|
cites | cdi_FETCH-LOGICAL-c372t-b0693cc10f586ca08dfb6b81f6ba59c249199f572c4d93465854b9c134fd619b3 |
container_end_page | |
container_issue | |
container_start_page | 115049 |
container_title | Expert systems with applications |
container_volume | 182 |
creator | Aras, Gizem Makaroğlu, Didem Demir, Seniz Cakir, Altan |
description | •Performances of LSTM and transformer-based networks are studied for Turkish NER.•A new transformer-based network with a CRF layer on top is introduced.•The state-of-the-art Turkish NER results are obtained with the proposed network
Named entity recognition (NER) is an extensively studied task that extracts and classifies named entities in a text. NER is crucial not only in downstream language processing applications such as relation extraction and question answering but also in large scale big data operations such as real-time analysis of online digital media content. Recent research efforts on Turkish, a less studied language with morphologically rich nature, have demonstrated the effectiveness of neural architectures on well-formed texts and yielded state-of-the art results by formulating the task as a sequence tagging problem. In this work, we empirically investigate the use of recent neural architectures (Bidirectional long short-term memory (BiLSTM) and Transformer-based networks) proposed for Turkish NER tagging in the same setting. Our results demonstrate that transformer-based networks which can model long-range context overcome the limitations of BiLSTM networks where different input features at the character, subword, and word levels are utilized. We also propose a transformer-based network with a conditional random field (CRF) layer that leads to the state-of-the-art result (95.95% f-measure) on a common dataset. Our study contributes to the literature that quantifies the impact of transfer learning on processing morphologically rich languages. |
doi_str_mv | 10.1016/j.eswa.2021.115049 |
format | article |
fullrecord | <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_2576366363</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S0957417421004905</els_id><sourcerecordid>2576366363</sourcerecordid><originalsourceid>FETCH-LOGICAL-c372t-b0693cc10f586ca08dfb6b81f6ba59c249199f572c4d93465854b9c134fd619b3</originalsourceid><addsrcrecordid>eNp9kE1LxDAQhoMouK7-AU8Bz61J0yYNeFkWv2DBy3rxEtI0qandZE3alf33pqxnYWAu88zM-wBwi1GOEab3fa7jj8wLVOAc4wqV_AwscM1IRhkn52CBeMWyErPyElzF2COEGUJsAT5WDuqDHCY5Wu-gNzBopd0InZ6CHGDU35N2SsNRdp11Hdz5Vg8RWge3U_iy8RM6udMtTIwdjzPtO2fnZdfgwsgh6pu_vgTvT4_b9Uu2eXt-Xa82mSKsGLMGUU6UwshUNVUS1a1paFNjQxtZcVWUHHNuKlaosuWkpFVdlQ1XmJSmpZg3ZAnuTnv3wadn4yh6PwWXToqiYpTQVCRNFacpFXyMQRuxD3Ynw1FgJGaHohezQzE7FCeHCXo4QSmyPlgdRFR21tHaFHQUrbf_4b-fDXrE</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2576366363</pqid></control><display><type>article</type><title>An evaluation of recent neural sequence tagging models in Turkish named entity recognition</title><source>Elsevier</source><creator>Aras, Gizem ; Makaroğlu, Didem ; Demir, Seniz ; Cakir, Altan</creator><creatorcontrib>Aras, Gizem ; Makaroğlu, Didem ; Demir, Seniz ; Cakir, Altan</creatorcontrib><description>•Performances of LSTM and transformer-based networks are studied for Turkish NER.•A new transformer-based network with a CRF layer on top is introduced.•The state-of-the-art Turkish NER results are obtained with the proposed network
Named entity recognition (NER) is an extensively studied task that extracts and classifies named entities in a text. NER is crucial not only in downstream language processing applications such as relation extraction and question answering but also in large scale big data operations such as real-time analysis of online digital media content. Recent research efforts on Turkish, a less studied language with morphologically rich nature, have demonstrated the effectiveness of neural architectures on well-formed texts and yielded state-of-the art results by formulating the task as a sequence tagging problem. In this work, we empirically investigate the use of recent neural architectures (Bidirectional long short-term memory (BiLSTM) and Transformer-based networks) proposed for Turkish NER tagging in the same setting. Our results demonstrate that transformer-based networks which can model long-range context overcome the limitations of BiLSTM networks where different input features at the character, subword, and word levels are utilized. We also propose a transformer-based network with a conditional random field (CRF) layer that leads to the state-of-the-art result (95.95% f-measure) on a common dataset. Our study contributes to the literature that quantifies the impact of transfer learning on processing morphologically rich languages.</description><identifier>ISSN: 0957-4174</identifier><identifier>EISSN: 1873-6793</identifier><identifier>DOI: 10.1016/j.eswa.2021.115049</identifier><language>eng</language><publisher>New York: Elsevier Ltd</publisher><subject>Conditional random fields ; CRF ; Digital media industry ; Marking ; Named entity recognition ; Natural language processing ; Networks ; Real time operation ; Recognition ; Transfer learning ; Transformers ; Turkish</subject><ispartof>Expert systems with applications, 2021-11, Vol.182, p.115049, Article 115049</ispartof><rights>2021 Elsevier Ltd</rights><rights>Copyright Elsevier BV Nov 15, 2021</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c372t-b0693cc10f586ca08dfb6b81f6ba59c249199f572c4d93465854b9c134fd619b3</citedby><cites>FETCH-LOGICAL-c372t-b0693cc10f586ca08dfb6b81f6ba59c249199f572c4d93465854b9c134fd619b3</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,776,780,27901,27902</link.rule.ids></links><search><creatorcontrib>Aras, Gizem</creatorcontrib><creatorcontrib>Makaroğlu, Didem</creatorcontrib><creatorcontrib>Demir, Seniz</creatorcontrib><creatorcontrib>Cakir, Altan</creatorcontrib><title>An evaluation of recent neural sequence tagging models in Turkish named entity recognition</title><title>Expert systems with applications</title><description>•Performances of LSTM and transformer-based networks are studied for Turkish NER.•A new transformer-based network with a CRF layer on top is introduced.•The state-of-the-art Turkish NER results are obtained with the proposed network
Named entity recognition (NER) is an extensively studied task that extracts and classifies named entities in a text. NER is crucial not only in downstream language processing applications such as relation extraction and question answering but also in large scale big data operations such as real-time analysis of online digital media content. Recent research efforts on Turkish, a less studied language with morphologically rich nature, have demonstrated the effectiveness of neural architectures on well-formed texts and yielded state-of-the art results by formulating the task as a sequence tagging problem. In this work, we empirically investigate the use of recent neural architectures (Bidirectional long short-term memory (BiLSTM) and Transformer-based networks) proposed for Turkish NER tagging in the same setting. Our results demonstrate that transformer-based networks which can model long-range context overcome the limitations of BiLSTM networks where different input features at the character, subword, and word levels are utilized. We also propose a transformer-based network with a conditional random field (CRF) layer that leads to the state-of-the-art result (95.95% f-measure) on a common dataset. Our study contributes to the literature that quantifies the impact of transfer learning on processing morphologically rich languages.</description><subject>Conditional random fields</subject><subject>CRF</subject><subject>Digital media industry</subject><subject>Marking</subject><subject>Named entity recognition</subject><subject>Natural language processing</subject><subject>Networks</subject><subject>Real time operation</subject><subject>Recognition</subject><subject>Transfer learning</subject><subject>Transformers</subject><subject>Turkish</subject><issn>0957-4174</issn><issn>1873-6793</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2021</creationdate><recordtype>article</recordtype><recordid>eNp9kE1LxDAQhoMouK7-AU8Bz61J0yYNeFkWv2DBy3rxEtI0qandZE3alf33pqxnYWAu88zM-wBwi1GOEab3fa7jj8wLVOAc4wqV_AwscM1IRhkn52CBeMWyErPyElzF2COEGUJsAT5WDuqDHCY5Wu-gNzBopd0InZ6CHGDU35N2SsNRdp11Hdz5Vg8RWge3U_iy8RM6udMtTIwdjzPtO2fnZdfgwsgh6pu_vgTvT4_b9Uu2eXt-Xa82mSKsGLMGUU6UwshUNVUS1a1paFNjQxtZcVWUHHNuKlaosuWkpFVdlQ1XmJSmpZg3ZAnuTnv3wadn4yh6PwWXToqiYpTQVCRNFacpFXyMQRuxD3Ynw1FgJGaHohezQzE7FCeHCXo4QSmyPlgdRFR21tHaFHQUrbf_4b-fDXrE</recordid><startdate>20211115</startdate><enddate>20211115</enddate><creator>Aras, Gizem</creator><creator>Makaroğlu, Didem</creator><creator>Demir, Seniz</creator><creator>Cakir, Altan</creator><general>Elsevier Ltd</general><general>Elsevier BV</general><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>8FD</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope></search><sort><creationdate>20211115</creationdate><title>An evaluation of recent neural sequence tagging models in Turkish named entity recognition</title><author>Aras, Gizem ; Makaroğlu, Didem ; Demir, Seniz ; Cakir, Altan</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c372t-b0693cc10f586ca08dfb6b81f6ba59c249199f572c4d93465854b9c134fd619b3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2021</creationdate><topic>Conditional random fields</topic><topic>CRF</topic><topic>Digital media industry</topic><topic>Marking</topic><topic>Named entity recognition</topic><topic>Natural language processing</topic><topic>Networks</topic><topic>Real time operation</topic><topic>Recognition</topic><topic>Transfer learning</topic><topic>Transformers</topic><topic>Turkish</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Aras, Gizem</creatorcontrib><creatorcontrib>Makaroğlu, Didem</creatorcontrib><creatorcontrib>Demir, Seniz</creatorcontrib><creatorcontrib>Cakir, Altan</creatorcontrib><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Technology Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><jtitle>Expert systems with applications</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Aras, Gizem</au><au>Makaroğlu, Didem</au><au>Demir, Seniz</au><au>Cakir, Altan</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>An evaluation of recent neural sequence tagging models in Turkish named entity recognition</atitle><jtitle>Expert systems with applications</jtitle><date>2021-11-15</date><risdate>2021</risdate><volume>182</volume><spage>115049</spage><pages>115049-</pages><artnum>115049</artnum><issn>0957-4174</issn><eissn>1873-6793</eissn><abstract>•Performances of LSTM and transformer-based networks are studied for Turkish NER.•A new transformer-based network with a CRF layer on top is introduced.•The state-of-the-art Turkish NER results are obtained with the proposed network
Named entity recognition (NER) is an extensively studied task that extracts and classifies named entities in a text. NER is crucial not only in downstream language processing applications such as relation extraction and question answering but also in large scale big data operations such as real-time analysis of online digital media content. Recent research efforts on Turkish, a less studied language with morphologically rich nature, have demonstrated the effectiveness of neural architectures on well-formed texts and yielded state-of-the art results by formulating the task as a sequence tagging problem. In this work, we empirically investigate the use of recent neural architectures (Bidirectional long short-term memory (BiLSTM) and Transformer-based networks) proposed for Turkish NER tagging in the same setting. Our results demonstrate that transformer-based networks which can model long-range context overcome the limitations of BiLSTM networks where different input features at the character, subword, and word levels are utilized. We also propose a transformer-based network with a conditional random field (CRF) layer that leads to the state-of-the-art result (95.95% f-measure) on a common dataset. Our study contributes to the literature that quantifies the impact of transfer learning on processing morphologically rich languages.</abstract><cop>New York</cop><pub>Elsevier Ltd</pub><doi>10.1016/j.eswa.2021.115049</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | ISSN: 0957-4174 |
ispartof | Expert systems with applications, 2021-11, Vol.182, p.115049, Article 115049 |
issn | 0957-4174 1873-6793 |
language | eng |
recordid | cdi_proquest_journals_2576366363 |
source | Elsevier |
subjects | Conditional random fields CRF Digital media industry Marking Named entity recognition Natural language processing Networks Real time operation Recognition Transfer learning Transformers Turkish |
title | An evaluation of recent neural sequence tagging models in Turkish named entity recognition |
url | http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-10T02%3A33%3A33IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=An%20evaluation%20of%20recent%20neural%20sequence%20tagging%20models%20in%20Turkish%20named%20entity%20recognition&rft.jtitle=Expert%20systems%20with%20applications&rft.au=Aras,%20Gizem&rft.date=2021-11-15&rft.volume=182&rft.spage=115049&rft.pages=115049-&rft.artnum=115049&rft.issn=0957-4174&rft.eissn=1873-6793&rft_id=info:doi/10.1016/j.eswa.2021.115049&rft_dat=%3Cproquest_cross%3E2576366363%3C/proquest_cross%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c372t-b0693cc10f586ca08dfb6b81f6ba59c249199f572c4d93465854b9c134fd619b3%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2576366363&rft_id=info:pmid/&rfr_iscdi=true |