Loading…
Constructing a Natural Language Inference Dataset using Generative Neural Networks
Natural Language Inference is an important task for Natural Language Understanding. It is concerned with classifying the logical relation between two sentences. In this paper, we propose several text generative neural networks for generating text hypothesis, which allows construction of new Natural...
Saved in:
Published in: | arXiv.org 2017-03 |
---|---|
Main Authors: | , |
Format: | Article |
Language: | English |
Subjects: | |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
cited_by | |
---|---|
cites | |
container_end_page | |
container_issue | |
container_start_page | |
container_title | arXiv.org |
container_volume | |
creator | Starc, Janez Mladenić, Dunja |
description | Natural Language Inference is an important task for Natural Language Understanding. It is concerned with classifying the logical relation between two sentences. In this paper, we propose several text generative neural networks for generating text hypothesis, which allows construction of new Natural Language Inference datasets. To evaluate the models, we propose a new metric -- the accuracy of the classifier trained on the generated dataset. The accuracy obtained by our best generative model is only 2.7% lower than the accuracy of the classifier trained on the original, human crafted dataset. Furthermore, the best generated dataset combined with the original dataset achieves the highest accuracy. The best model learns a mapping embedding for each training example. By comparing various metrics we show that datasets that obtain higher ROUGE or METEOR scores do not necessarily yield higher classification accuracies. We also provide analysis of what are the characteristics of a good dataset including the distinguishability of the generated datasets from the original one. |
format | article |
fullrecord | <record><control><sourceid>proquest</sourceid><recordid>TN_cdi_proquest_journals_2075296068</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2075296068</sourcerecordid><originalsourceid>FETCH-proquest_journals_20752960683</originalsourceid><addsrcrecordid>eNqNi8sKwjAQAIMgKNp_CHguxMQ-PNcnSA_ivSxlW1pLotmN_r4P_ABPc5iZkZhqY5ZxvtJ6IiKiXiml00wniZmKc-EssQ81d7aVIEvg4GGQJ7BtgBbl0Tbo0dYoN8BAyDLQJ92jRQ_cPVCW-F1K5KfzV5qLcQMDYfTjTCx220txiG_e3QMSV70L3r5VpVWW6HWq0tz8V70AC0Q_-Q</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2075296068</pqid></control><display><type>article</type><title>Constructing a Natural Language Inference Dataset using Generative Neural Networks</title><source>Publicly Available Content Database</source><creator>Starc, Janez ; Mladenić, Dunja</creator><creatorcontrib>Starc, Janez ; Mladenić, Dunja</creatorcontrib><description>Natural Language Inference is an important task for Natural Language Understanding. It is concerned with classifying the logical relation between two sentences. In this paper, we propose several text generative neural networks for generating text hypothesis, which allows construction of new Natural Language Inference datasets. To evaluate the models, we propose a new metric -- the accuracy of the classifier trained on the generated dataset. The accuracy obtained by our best generative model is only 2.7% lower than the accuracy of the classifier trained on the original, human crafted dataset. Furthermore, the best generated dataset combined with the original dataset achieves the highest accuracy. The best model learns a mapping embedding for each training example. By comparing various metrics we show that datasets that obtain higher ROUGE or METEOR scores do not necessarily yield higher classification accuracies. We also provide analysis of what are the characteristics of a good dataset including the distinguishability of the generated datasets from the original one.</description><identifier>EISSN: 2331-8422</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Accuracy ; Classification ; Classifiers ; Datasets ; Inference ; Mapping ; Model accuracy ; Natural language ; Neural networks ; Sentences</subject><ispartof>arXiv.org, 2017-03</ispartof><rights>2017. This work is published under http://arxiv.org/licenses/nonexclusive-distrib/1.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://www.proquest.com/docview/2075296068?pq-origsite=primo$$EHTML$$P50$$Gproquest$$Hfree_for_read</linktohtml><link.rule.ids>780,784,25753,37012,44590</link.rule.ids></links><search><creatorcontrib>Starc, Janez</creatorcontrib><creatorcontrib>Mladenić, Dunja</creatorcontrib><title>Constructing a Natural Language Inference Dataset using Generative Neural Networks</title><title>arXiv.org</title><description>Natural Language Inference is an important task for Natural Language Understanding. It is concerned with classifying the logical relation between two sentences. In this paper, we propose several text generative neural networks for generating text hypothesis, which allows construction of new Natural Language Inference datasets. To evaluate the models, we propose a new metric -- the accuracy of the classifier trained on the generated dataset. The accuracy obtained by our best generative model is only 2.7% lower than the accuracy of the classifier trained on the original, human crafted dataset. Furthermore, the best generated dataset combined with the original dataset achieves the highest accuracy. The best model learns a mapping embedding for each training example. By comparing various metrics we show that datasets that obtain higher ROUGE or METEOR scores do not necessarily yield higher classification accuracies. We also provide analysis of what are the characteristics of a good dataset including the distinguishability of the generated datasets from the original one.</description><subject>Accuracy</subject><subject>Classification</subject><subject>Classifiers</subject><subject>Datasets</subject><subject>Inference</subject><subject>Mapping</subject><subject>Model accuracy</subject><subject>Natural language</subject><subject>Neural networks</subject><subject>Sentences</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2017</creationdate><recordtype>article</recordtype><sourceid>PIMPY</sourceid><recordid>eNqNi8sKwjAQAIMgKNp_CHguxMQ-PNcnSA_ivSxlW1pLotmN_r4P_ABPc5iZkZhqY5ZxvtJ6IiKiXiml00wniZmKc-EssQ81d7aVIEvg4GGQJ7BtgBbl0Tbo0dYoN8BAyDLQJ92jRQ_cPVCW-F1K5KfzV5qLcQMDYfTjTCx220txiG_e3QMSV70L3r5VpVWW6HWq0tz8V70AC0Q_-Q</recordid><startdate>20170327</startdate><enddate>20170327</enddate><creator>Starc, Janez</creator><creator>Mladenić, Dunja</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope></search><sort><creationdate>20170327</creationdate><title>Constructing a Natural Language Inference Dataset using Generative Neural Networks</title><author>Starc, Janez ; Mladenić, Dunja</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-proquest_journals_20752960683</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2017</creationdate><topic>Accuracy</topic><topic>Classification</topic><topic>Classifiers</topic><topic>Datasets</topic><topic>Inference</topic><topic>Mapping</topic><topic>Model accuracy</topic><topic>Natural language</topic><topic>Neural networks</topic><topic>Sentences</topic><toplevel>online_resources</toplevel><creatorcontrib>Starc, Janez</creatorcontrib><creatorcontrib>Mladenić, Dunja</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science & Engineering Collection</collection><collection>ProQuest Central (Alumni)</collection><collection>ProQuest Central</collection><collection>ProQuest Central Essentials</collection><collection>AUTh Library subscriptions: ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering Collection</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Starc, Janez</au><au>Mladenić, Dunja</au><format>book</format><genre>document</genre><ristype>GEN</ristype><atitle>Constructing a Natural Language Inference Dataset using Generative Neural Networks</atitle><jtitle>arXiv.org</jtitle><date>2017-03-27</date><risdate>2017</risdate><eissn>2331-8422</eissn><abstract>Natural Language Inference is an important task for Natural Language Understanding. It is concerned with classifying the logical relation between two sentences. In this paper, we propose several text generative neural networks for generating text hypothesis, which allows construction of new Natural Language Inference datasets. To evaluate the models, we propose a new metric -- the accuracy of the classifier trained on the generated dataset. The accuracy obtained by our best generative model is only 2.7% lower than the accuracy of the classifier trained on the original, human crafted dataset. Furthermore, the best generated dataset combined with the original dataset achieves the highest accuracy. The best model learns a mapping embedding for each training example. By comparing various metrics we show that datasets that obtain higher ROUGE or METEOR scores do not necessarily yield higher classification accuracies. We also provide analysis of what are the characteristics of a good dataset including the distinguishability of the generated datasets from the original one.</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | EISSN: 2331-8422 |
ispartof | arXiv.org, 2017-03 |
issn | 2331-8422 |
language | eng |
recordid | cdi_proquest_journals_2075296068 |
source | Publicly Available Content Database |
subjects | Accuracy Classification Classifiers Datasets Inference Mapping Model accuracy Natural language Neural networks Sentences |
title | Constructing a Natural Language Inference Dataset using Generative Neural Networks |
url | http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-06T05%3A27%3A19IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=document&rft.atitle=Constructing%20a%20Natural%20Language%20Inference%20Dataset%20using%20Generative%20Neural%20Networks&rft.jtitle=arXiv.org&rft.au=Starc,%20Janez&rft.date=2017-03-27&rft.eissn=2331-8422&rft_id=info:doi/&rft_dat=%3Cproquest%3E2075296068%3C/proquest%3E%3Cgrp_id%3Ecdi_FETCH-proquest_journals_20752960683%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2075296068&rft_id=info:pmid/&rfr_iscdi=true |