Loading…
Generation of Coherent Multi-Sentence Texts with a Coherence Mechanism
Automatic generation of long texts containing multiple sentences has many applications in the field of Natural Language Processing (NLP) including question answering, machine translation, and paraphrase generation, etc. However, in terms of readability, the long texts generated by machines are not c...
Saved in:
Published in: | Computer speech & language 2023-03, Vol.78, p.101457, Article 101457 |
---|---|
Main Authors: | , , , , |
Format: | Article |
Language: | English |
Subjects: | |
Citations: | Items that this one cites Items that cite this one |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
cited_by | cdi_FETCH-LOGICAL-c297t-be448a9cc0532f91493b9331ca441b48f41efa167af53857e02df72f7ec0195f3 |
---|---|
cites | cdi_FETCH-LOGICAL-c297t-be448a9cc0532f91493b9331ca441b48f41efa167af53857e02df72f7ec0195f3 |
container_end_page | |
container_issue | |
container_start_page | 101457 |
container_title | Computer speech & language |
container_volume | 78 |
creator | Zhao, Qingjuan Niu, Jianwei Liu, Xuefeng He, Wenbo Tang, Shaojie |
description | Automatic generation of long texts containing multiple sentences has many applications in the field of Natural Language Processing (NLP) including question answering, machine translation, and paraphrase generation, etc. However, in terms of readability, the long texts generated by machines are not comparable to those organized by human beings. Through statistics, we observed that human-organized texts generally have a special property: one or more of the words (particularly nouns and pronouns) appeared in one sentence will reappear in the next one in the same or a different form. This repetition of words in consecutive sentences can greatly improve the readability. Based on this observation, we propose CMST, a deep neural network model for generating Coherent Multi-Sentence Texts. CMST explicitly incorporates a training strategy of coherence mechanism to evaluate the repetition of words in consecutive sentences. We evaluate the performance of the CMST on the CNN/Daily Mail dataset. The experimental results show that, compared with the baseline models, CMST not only improves the readability of the generated texts, but achieves higher METEOR and ROUGE values.
•The reappearance of words in adjacent sentences could make the text read coherently.•The coherence level is defined and incorporated into the objective function.•The results show that the generated texts achieve higher evaluation scores. |
doi_str_mv | 10.1016/j.csl.2022.101457 |
format | article |
fullrecord | <record><control><sourceid>elsevier_cross</sourceid><recordid>TN_cdi_crossref_primary_10_1016_j_csl_2022_101457</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S0885230822000808</els_id><sourcerecordid>S0885230822000808</sourcerecordid><originalsourceid>FETCH-LOGICAL-c297t-be448a9cc0532f91493b9331ca441b48f41efa167af53857e02df72f7ec0195f3</originalsourceid><addsrcrecordid>eNp9kMtKAzEUhoMoWKsP4G5eYOrJbZLgSopthRYX1nVI0xOa0s5IEm9v7wzVratz4XyHn4-QWwoTCrS52098PkwYMDbMQqozMqJgZK15w8_JCLSWNeOgL8lVznsAaKRQIzKbY4vJldi1VReqabfDhG2pVu-HEuuXvsXWY7XGr5Krz1h2lfs76tcr9DvXxny8JhfBHTLe_NYxeZ09rqeLevk8f5o-LGvPjCr1BoXQzngPkrNgqDB8Yzin3glBN0IHQTE42igXJNdSIbBtUCwo9ECNDHxM6OmvT13OCYN9S_Ho0relYAcRdm97EXYQYU8ieub-xGAf7CNistnHIf42JvTFbrv4D_0DYZRlnw</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Generation of Coherent Multi-Sentence Texts with a Coherence Mechanism</title><source>ScienceDirect Freedom Collection 2022-2024</source><creator>Zhao, Qingjuan ; Niu, Jianwei ; Liu, Xuefeng ; He, Wenbo ; Tang, Shaojie</creator><creatorcontrib>Zhao, Qingjuan ; Niu, Jianwei ; Liu, Xuefeng ; He, Wenbo ; Tang, Shaojie</creatorcontrib><description>Automatic generation of long texts containing multiple sentences has many applications in the field of Natural Language Processing (NLP) including question answering, machine translation, and paraphrase generation, etc. However, in terms of readability, the long texts generated by machines are not comparable to those organized by human beings. Through statistics, we observed that human-organized texts generally have a special property: one or more of the words (particularly nouns and pronouns) appeared in one sentence will reappear in the next one in the same or a different form. This repetition of words in consecutive sentences can greatly improve the readability. Based on this observation, we propose CMST, a deep neural network model for generating Coherent Multi-Sentence Texts. CMST explicitly incorporates a training strategy of coherence mechanism to evaluate the repetition of words in consecutive sentences. We evaluate the performance of the CMST on the CNN/Daily Mail dataset. The experimental results show that, compared with the baseline models, CMST not only improves the readability of the generated texts, but achieves higher METEOR and ROUGE values.
•The reappearance of words in adjacent sentences could make the text read coherently.•The coherence level is defined and incorporated into the objective function.•The results show that the generated texts achieve higher evaluation scores.</description><identifier>ISSN: 0885-2308</identifier><identifier>EISSN: 1095-8363</identifier><identifier>DOI: 10.1016/j.csl.2022.101457</identifier><language>eng</language><publisher>Elsevier Ltd</publisher><subject>Abstractive summarization ; Attention mechanism ; Coherence mechanism ; Multiple sentences summarization ; Text generation</subject><ispartof>Computer speech & language, 2023-03, Vol.78, p.101457, Article 101457</ispartof><rights>2022 Elsevier Ltd</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c297t-be448a9cc0532f91493b9331ca441b48f41efa167af53857e02df72f7ec0195f3</citedby><cites>FETCH-LOGICAL-c297t-be448a9cc0532f91493b9331ca441b48f41efa167af53857e02df72f7ec0195f3</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,780,784,27924,27925</link.rule.ids></links><search><creatorcontrib>Zhao, Qingjuan</creatorcontrib><creatorcontrib>Niu, Jianwei</creatorcontrib><creatorcontrib>Liu, Xuefeng</creatorcontrib><creatorcontrib>He, Wenbo</creatorcontrib><creatorcontrib>Tang, Shaojie</creatorcontrib><title>Generation of Coherent Multi-Sentence Texts with a Coherence Mechanism</title><title>Computer speech & language</title><description>Automatic generation of long texts containing multiple sentences has many applications in the field of Natural Language Processing (NLP) including question answering, machine translation, and paraphrase generation, etc. However, in terms of readability, the long texts generated by machines are not comparable to those organized by human beings. Through statistics, we observed that human-organized texts generally have a special property: one or more of the words (particularly nouns and pronouns) appeared in one sentence will reappear in the next one in the same or a different form. This repetition of words in consecutive sentences can greatly improve the readability. Based on this observation, we propose CMST, a deep neural network model for generating Coherent Multi-Sentence Texts. CMST explicitly incorporates a training strategy of coherence mechanism to evaluate the repetition of words in consecutive sentences. We evaluate the performance of the CMST on the CNN/Daily Mail dataset. The experimental results show that, compared with the baseline models, CMST not only improves the readability of the generated texts, but achieves higher METEOR and ROUGE values.
•The reappearance of words in adjacent sentences could make the text read coherently.•The coherence level is defined and incorporated into the objective function.•The results show that the generated texts achieve higher evaluation scores.</description><subject>Abstractive summarization</subject><subject>Attention mechanism</subject><subject>Coherence mechanism</subject><subject>Multiple sentences summarization</subject><subject>Text generation</subject><issn>0885-2308</issn><issn>1095-8363</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><recordid>eNp9kMtKAzEUhoMoWKsP4G5eYOrJbZLgSopthRYX1nVI0xOa0s5IEm9v7wzVratz4XyHn4-QWwoTCrS52098PkwYMDbMQqozMqJgZK15w8_JCLSWNeOgL8lVznsAaKRQIzKbY4vJldi1VReqabfDhG2pVu-HEuuXvsXWY7XGr5Krz1h2lfs76tcr9DvXxny8JhfBHTLe_NYxeZ09rqeLevk8f5o-LGvPjCr1BoXQzngPkrNgqDB8Yzin3glBN0IHQTE42igXJNdSIbBtUCwo9ECNDHxM6OmvT13OCYN9S_Ho0relYAcRdm97EXYQYU8ieub-xGAf7CNistnHIf42JvTFbrv4D_0DYZRlnw</recordid><startdate>202303</startdate><enddate>202303</enddate><creator>Zhao, Qingjuan</creator><creator>Niu, Jianwei</creator><creator>Liu, Xuefeng</creator><creator>He, Wenbo</creator><creator>Tang, Shaojie</creator><general>Elsevier Ltd</general><scope>AAYXX</scope><scope>CITATION</scope></search><sort><creationdate>202303</creationdate><title>Generation of Coherent Multi-Sentence Texts with a Coherence Mechanism</title><author>Zhao, Qingjuan ; Niu, Jianwei ; Liu, Xuefeng ; He, Wenbo ; Tang, Shaojie</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c297t-be448a9cc0532f91493b9331ca441b48f41efa167af53857e02df72f7ec0195f3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Abstractive summarization</topic><topic>Attention mechanism</topic><topic>Coherence mechanism</topic><topic>Multiple sentences summarization</topic><topic>Text generation</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Zhao, Qingjuan</creatorcontrib><creatorcontrib>Niu, Jianwei</creatorcontrib><creatorcontrib>Liu, Xuefeng</creatorcontrib><creatorcontrib>He, Wenbo</creatorcontrib><creatorcontrib>Tang, Shaojie</creatorcontrib><collection>CrossRef</collection><jtitle>Computer speech & language</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Zhao, Qingjuan</au><au>Niu, Jianwei</au><au>Liu, Xuefeng</au><au>He, Wenbo</au><au>Tang, Shaojie</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Generation of Coherent Multi-Sentence Texts with a Coherence Mechanism</atitle><jtitle>Computer speech & language</jtitle><date>2023-03</date><risdate>2023</risdate><volume>78</volume><spage>101457</spage><pages>101457-</pages><artnum>101457</artnum><issn>0885-2308</issn><eissn>1095-8363</eissn><abstract>Automatic generation of long texts containing multiple sentences has many applications in the field of Natural Language Processing (NLP) including question answering, machine translation, and paraphrase generation, etc. However, in terms of readability, the long texts generated by machines are not comparable to those organized by human beings. Through statistics, we observed that human-organized texts generally have a special property: one or more of the words (particularly nouns and pronouns) appeared in one sentence will reappear in the next one in the same or a different form. This repetition of words in consecutive sentences can greatly improve the readability. Based on this observation, we propose CMST, a deep neural network model for generating Coherent Multi-Sentence Texts. CMST explicitly incorporates a training strategy of coherence mechanism to evaluate the repetition of words in consecutive sentences. We evaluate the performance of the CMST on the CNN/Daily Mail dataset. The experimental results show that, compared with the baseline models, CMST not only improves the readability of the generated texts, but achieves higher METEOR and ROUGE values.
•The reappearance of words in adjacent sentences could make the text read coherently.•The coherence level is defined and incorporated into the objective function.•The results show that the generated texts achieve higher evaluation scores.</abstract><pub>Elsevier Ltd</pub><doi>10.1016/j.csl.2022.101457</doi></addata></record> |
fulltext | fulltext |
identifier | ISSN: 0885-2308 |
ispartof | Computer speech & language, 2023-03, Vol.78, p.101457, Article 101457 |
issn | 0885-2308 1095-8363 |
language | eng |
recordid | cdi_crossref_primary_10_1016_j_csl_2022_101457 |
source | ScienceDirect Freedom Collection 2022-2024 |
subjects | Abstractive summarization Attention mechanism Coherence mechanism Multiple sentences summarization Text generation |
title | Generation of Coherent Multi-Sentence Texts with a Coherence Mechanism |
url | http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-05T06%3A21%3A40IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-elsevier_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Generation%20of%20Coherent%20Multi-Sentence%20Texts%20with%20a%20Coherence%20Mechanism&rft.jtitle=Computer%20speech%20&%20language&rft.au=Zhao,%20Qingjuan&rft.date=2023-03&rft.volume=78&rft.spage=101457&rft.pages=101457-&rft.artnum=101457&rft.issn=0885-2308&rft.eissn=1095-8363&rft_id=info:doi/10.1016/j.csl.2022.101457&rft_dat=%3Celsevier_cross%3ES0885230822000808%3C/elsevier_cross%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c297t-be448a9cc0532f91493b9331ca441b48f41efa167af53857e02df72f7ec0195f3%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |