Loading…

An evaluation of recent neural sequence tagging models in Turkish named entity recognition

•Performances of LSTM and transformer-based networks are studied for Turkish NER.•A new transformer-based network with a CRF layer on top is introduced.•The state-of-the-art Turkish NER results are obtained with the proposed network Named entity recognition (NER) is an extensively studied task that...

Full description

Saved in:
Bibliographic Details
Published in:Expert systems with applications 2021-11, Vol.182, p.115049, Article 115049
Main Authors: Aras, Gizem, Makaroğlu, Didem, Demir, Seniz, Cakir, Altan
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Items that cite this one
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by cdi_FETCH-LOGICAL-c372t-b0693cc10f586ca08dfb6b81f6ba59c249199f572c4d93465854b9c134fd619b3
cites cdi_FETCH-LOGICAL-c372t-b0693cc10f586ca08dfb6b81f6ba59c249199f572c4d93465854b9c134fd619b3
container_end_page
container_issue
container_start_page 115049
container_title Expert systems with applications
container_volume 182
creator Aras, Gizem
Makaroğlu, Didem
Demir, Seniz
Cakir, Altan
description •Performances of LSTM and transformer-based networks are studied for Turkish NER.•A new transformer-based network with a CRF layer on top is introduced.•The state-of-the-art Turkish NER results are obtained with the proposed network Named entity recognition (NER) is an extensively studied task that extracts and classifies named entities in a text. NER is crucial not only in downstream language processing applications such as relation extraction and question answering but also in large scale big data operations such as real-time analysis of online digital media content. Recent research efforts on Turkish, a less studied language with morphologically rich nature, have demonstrated the effectiveness of neural architectures on well-formed texts and yielded state-of-the art results by formulating the task as a sequence tagging problem. In this work, we empirically investigate the use of recent neural architectures (Bidirectional long short-term memory (BiLSTM) and Transformer-based networks) proposed for Turkish NER tagging in the same setting. Our results demonstrate that transformer-based networks which can model long-range context overcome the limitations of BiLSTM networks where different input features at the character, subword, and word levels are utilized. We also propose a transformer-based network with a conditional random field (CRF) layer that leads to the state-of-the-art result (95.95% f-measure) on a common dataset. Our study contributes to the literature that quantifies the impact of transfer learning on processing morphologically rich languages.
doi_str_mv 10.1016/j.eswa.2021.115049
format article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_2576366363</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S0957417421004905</els_id><sourcerecordid>2576366363</sourcerecordid><originalsourceid>FETCH-LOGICAL-c372t-b0693cc10f586ca08dfb6b81f6ba59c249199f572c4d93465854b9c134fd619b3</originalsourceid><addsrcrecordid>eNp9kE1LxDAQhoMouK7-AU8Bz61J0yYNeFkWv2DBy3rxEtI0qandZE3alf33pqxnYWAu88zM-wBwi1GOEab3fa7jj8wLVOAc4wqV_AwscM1IRhkn52CBeMWyErPyElzF2COEGUJsAT5WDuqDHCY5Wu-gNzBopd0InZ6CHGDU35N2SsNRdp11Hdz5Vg8RWge3U_iy8RM6udMtTIwdjzPtO2fnZdfgwsgh6pu_vgTvT4_b9Uu2eXt-Xa82mSKsGLMGUU6UwshUNVUS1a1paFNjQxtZcVWUHHNuKlaosuWkpFVdlQ1XmJSmpZg3ZAnuTnv3wadn4yh6PwWXToqiYpTQVCRNFacpFXyMQRuxD3Ynw1FgJGaHohezQzE7FCeHCXo4QSmyPlgdRFR21tHaFHQUrbf_4b-fDXrE</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2576366363</pqid></control><display><type>article</type><title>An evaluation of recent neural sequence tagging models in Turkish named entity recognition</title><source>Elsevier</source><creator>Aras, Gizem ; Makaroğlu, Didem ; Demir, Seniz ; Cakir, Altan</creator><creatorcontrib>Aras, Gizem ; Makaroğlu, Didem ; Demir, Seniz ; Cakir, Altan</creatorcontrib><description>•Performances of LSTM and transformer-based networks are studied for Turkish NER.•A new transformer-based network with a CRF layer on top is introduced.•The state-of-the-art Turkish NER results are obtained with the proposed network Named entity recognition (NER) is an extensively studied task that extracts and classifies named entities in a text. NER is crucial not only in downstream language processing applications such as relation extraction and question answering but also in large scale big data operations such as real-time analysis of online digital media content. Recent research efforts on Turkish, a less studied language with morphologically rich nature, have demonstrated the effectiveness of neural architectures on well-formed texts and yielded state-of-the art results by formulating the task as a sequence tagging problem. In this work, we empirically investigate the use of recent neural architectures (Bidirectional long short-term memory (BiLSTM) and Transformer-based networks) proposed for Turkish NER tagging in the same setting. Our results demonstrate that transformer-based networks which can model long-range context overcome the limitations of BiLSTM networks where different input features at the character, subword, and word levels are utilized. We also propose a transformer-based network with a conditional random field (CRF) layer that leads to the state-of-the-art result (95.95% f-measure) on a common dataset. Our study contributes to the literature that quantifies the impact of transfer learning on processing morphologically rich languages.</description><identifier>ISSN: 0957-4174</identifier><identifier>EISSN: 1873-6793</identifier><identifier>DOI: 10.1016/j.eswa.2021.115049</identifier><language>eng</language><publisher>New York: Elsevier Ltd</publisher><subject>Conditional random fields ; CRF ; Digital media industry ; Marking ; Named entity recognition ; Natural language processing ; Networks ; Real time operation ; Recognition ; Transfer learning ; Transformers ; Turkish</subject><ispartof>Expert systems with applications, 2021-11, Vol.182, p.115049, Article 115049</ispartof><rights>2021 Elsevier Ltd</rights><rights>Copyright Elsevier BV Nov 15, 2021</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c372t-b0693cc10f586ca08dfb6b81f6ba59c249199f572c4d93465854b9c134fd619b3</citedby><cites>FETCH-LOGICAL-c372t-b0693cc10f586ca08dfb6b81f6ba59c249199f572c4d93465854b9c134fd619b3</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,776,780,27901,27902</link.rule.ids></links><search><creatorcontrib>Aras, Gizem</creatorcontrib><creatorcontrib>Makaroğlu, Didem</creatorcontrib><creatorcontrib>Demir, Seniz</creatorcontrib><creatorcontrib>Cakir, Altan</creatorcontrib><title>An evaluation of recent neural sequence tagging models in Turkish named entity recognition</title><title>Expert systems with applications</title><description>•Performances of LSTM and transformer-based networks are studied for Turkish NER.•A new transformer-based network with a CRF layer on top is introduced.•The state-of-the-art Turkish NER results are obtained with the proposed network Named entity recognition (NER) is an extensively studied task that extracts and classifies named entities in a text. NER is crucial not only in downstream language processing applications such as relation extraction and question answering but also in large scale big data operations such as real-time analysis of online digital media content. Recent research efforts on Turkish, a less studied language with morphologically rich nature, have demonstrated the effectiveness of neural architectures on well-formed texts and yielded state-of-the art results by formulating the task as a sequence tagging problem. In this work, we empirically investigate the use of recent neural architectures (Bidirectional long short-term memory (BiLSTM) and Transformer-based networks) proposed for Turkish NER tagging in the same setting. Our results demonstrate that transformer-based networks which can model long-range context overcome the limitations of BiLSTM networks where different input features at the character, subword, and word levels are utilized. We also propose a transformer-based network with a conditional random field (CRF) layer that leads to the state-of-the-art result (95.95% f-measure) on a common dataset. Our study contributes to the literature that quantifies the impact of transfer learning on processing morphologically rich languages.</description><subject>Conditional random fields</subject><subject>CRF</subject><subject>Digital media industry</subject><subject>Marking</subject><subject>Named entity recognition</subject><subject>Natural language processing</subject><subject>Networks</subject><subject>Real time operation</subject><subject>Recognition</subject><subject>Transfer learning</subject><subject>Transformers</subject><subject>Turkish</subject><issn>0957-4174</issn><issn>1873-6793</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2021</creationdate><recordtype>article</recordtype><recordid>eNp9kE1LxDAQhoMouK7-AU8Bz61J0yYNeFkWv2DBy3rxEtI0qandZE3alf33pqxnYWAu88zM-wBwi1GOEab3fa7jj8wLVOAc4wqV_AwscM1IRhkn52CBeMWyErPyElzF2COEGUJsAT5WDuqDHCY5Wu-gNzBopd0InZ6CHGDU35N2SsNRdp11Hdz5Vg8RWge3U_iy8RM6udMtTIwdjzPtO2fnZdfgwsgh6pu_vgTvT4_b9Uu2eXt-Xa82mSKsGLMGUU6UwshUNVUS1a1paFNjQxtZcVWUHHNuKlaosuWkpFVdlQ1XmJSmpZg3ZAnuTnv3wadn4yh6PwWXToqiYpTQVCRNFacpFXyMQRuxD3Ynw1FgJGaHohezQzE7FCeHCXo4QSmyPlgdRFR21tHaFHQUrbf_4b-fDXrE</recordid><startdate>20211115</startdate><enddate>20211115</enddate><creator>Aras, Gizem</creator><creator>Makaroğlu, Didem</creator><creator>Demir, Seniz</creator><creator>Cakir, Altan</creator><general>Elsevier Ltd</general><general>Elsevier BV</general><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>8FD</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope></search><sort><creationdate>20211115</creationdate><title>An evaluation of recent neural sequence tagging models in Turkish named entity recognition</title><author>Aras, Gizem ; Makaroğlu, Didem ; Demir, Seniz ; Cakir, Altan</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c372t-b0693cc10f586ca08dfb6b81f6ba59c249199f572c4d93465854b9c134fd619b3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2021</creationdate><topic>Conditional random fields</topic><topic>CRF</topic><topic>Digital media industry</topic><topic>Marking</topic><topic>Named entity recognition</topic><topic>Natural language processing</topic><topic>Networks</topic><topic>Real time operation</topic><topic>Recognition</topic><topic>Transfer learning</topic><topic>Transformers</topic><topic>Turkish</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Aras, Gizem</creatorcontrib><creatorcontrib>Makaroğlu, Didem</creatorcontrib><creatorcontrib>Demir, Seniz</creatorcontrib><creatorcontrib>Cakir, Altan</creatorcontrib><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Technology Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><jtitle>Expert systems with applications</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Aras, Gizem</au><au>Makaroğlu, Didem</au><au>Demir, Seniz</au><au>Cakir, Altan</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>An evaluation of recent neural sequence tagging models in Turkish named entity recognition</atitle><jtitle>Expert systems with applications</jtitle><date>2021-11-15</date><risdate>2021</risdate><volume>182</volume><spage>115049</spage><pages>115049-</pages><artnum>115049</artnum><issn>0957-4174</issn><eissn>1873-6793</eissn><abstract>•Performances of LSTM and transformer-based networks are studied for Turkish NER.•A new transformer-based network with a CRF layer on top is introduced.•The state-of-the-art Turkish NER results are obtained with the proposed network Named entity recognition (NER) is an extensively studied task that extracts and classifies named entities in a text. NER is crucial not only in downstream language processing applications such as relation extraction and question answering but also in large scale big data operations such as real-time analysis of online digital media content. Recent research efforts on Turkish, a less studied language with morphologically rich nature, have demonstrated the effectiveness of neural architectures on well-formed texts and yielded state-of-the art results by formulating the task as a sequence tagging problem. In this work, we empirically investigate the use of recent neural architectures (Bidirectional long short-term memory (BiLSTM) and Transformer-based networks) proposed for Turkish NER tagging in the same setting. Our results demonstrate that transformer-based networks which can model long-range context overcome the limitations of BiLSTM networks where different input features at the character, subword, and word levels are utilized. We also propose a transformer-based network with a conditional random field (CRF) layer that leads to the state-of-the-art result (95.95% f-measure) on a common dataset. Our study contributes to the literature that quantifies the impact of transfer learning on processing morphologically rich languages.</abstract><cop>New York</cop><pub>Elsevier Ltd</pub><doi>10.1016/j.eswa.2021.115049</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 0957-4174
ispartof Expert systems with applications, 2021-11, Vol.182, p.115049, Article 115049
issn 0957-4174
1873-6793
language eng
recordid cdi_proquest_journals_2576366363
source Elsevier
subjects Conditional random fields
CRF
Digital media industry
Marking
Named entity recognition
Natural language processing
Networks
Real time operation
Recognition
Transfer learning
Transformers
Turkish
title An evaluation of recent neural sequence tagging models in Turkish named entity recognition
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-10T02%3A33%3A33IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=An%20evaluation%20of%20recent%20neural%20sequence%20tagging%20models%20in%20Turkish%20named%20entity%20recognition&rft.jtitle=Expert%20systems%20with%20applications&rft.au=Aras,%20Gizem&rft.date=2021-11-15&rft.volume=182&rft.spage=115049&rft.pages=115049-&rft.artnum=115049&rft.issn=0957-4174&rft.eissn=1873-6793&rft_id=info:doi/10.1016/j.eswa.2021.115049&rft_dat=%3Cproquest_cross%3E2576366363%3C/proquest_cross%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c372t-b0693cc10f586ca08dfb6b81f6ba59c249199f572c4d93465854b9c134fd619b3%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2576366363&rft_id=info:pmid/&rfr_iscdi=true