Loading…

Enhancing Automated Scoring of Math Self-Explanation Quality Using LLM-Generated Datasets: A Semi-Supervised Approach

In the realm of mathematics education, self-explanation stands as a crucial learning mechanism, allowing learners to articulate their comprehension of intricate mathematical concepts and strategies. As digital learning platforms grow in prominence, there are mounting opportunities to collect and uti...

Full description

Saved in:
Bibliographic Details
Published in:Computers (Basel) 2023-11, Vol.12 (11), p.217
Main Authors: Nakamoto, Ryosuke, Flanagan, Brendan, Yamauchi, Taisei, Dai, Yiling, Takami, Kyosuke, Ogata, Hiroaki
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Items that cite this one
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by cdi_FETCH-LOGICAL-c421t-5b327573996644e0da2999e30ee1be9ffac72ecdbd208670fd13479563bc0b193
cites cdi_FETCH-LOGICAL-c421t-5b327573996644e0da2999e30ee1be9ffac72ecdbd208670fd13479563bc0b193
container_end_page
container_issue 11
container_start_page 217
container_title Computers (Basel)
container_volume 12
creator Nakamoto, Ryosuke
Flanagan, Brendan
Yamauchi, Taisei
Dai, Yiling
Takami, Kyosuke
Ogata, Hiroaki
description In the realm of mathematics education, self-explanation stands as a crucial learning mechanism, allowing learners to articulate their comprehension of intricate mathematical concepts and strategies. As digital learning platforms grow in prominence, there are mounting opportunities to collect and utilize mathematical self-explanations. However, these opportunities are met with challenges in automated evaluation. Automatic scoring of mathematical self-explanations is crucial for preprocessing tasks, including the categorization of learner responses, identification of common misconceptions, and the creation of tailored feedback and model solutions. Nevertheless, this task is hindered by the dearth of ample sample sets. Our research introduces a semi-supervised technique using the large language model (LLM), specifically its Japanese variant, to enrich datasets for the automated scoring of mathematical self-explanations. We rigorously evaluated the quality of self-explanations across five datasets, ranging from human-evaluated originals to ones devoid of original content. Our results show that combining LLM-based explanations with mathematical material significantly improves the model’s accuracy. Interestingly, there is an optimal limit to how many synthetic self-explanation data can benefit the system. Exceeding this limit does not further improve outcomes. This study thus highlights the need for careful consideration when integrating synthetic data into solutions, especially within the mathematics discipline.
doi_str_mv 10.3390/computers12110217
format article
fullrecord <record><control><sourceid>gale_doaj_</sourceid><recordid>TN_cdi_doaj_primary_oai_doaj_org_article_ea5bb076f62045a78e6f9ab1933a816d</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><galeid>A774318956</galeid><doaj_id>oai_doaj_org_article_ea5bb076f62045a78e6f9ab1933a816d</doaj_id><sourcerecordid>A774318956</sourcerecordid><originalsourceid>FETCH-LOGICAL-c421t-5b327573996644e0da2999e30ee1be9ffac72ecdbd208670fd13479563bc0b193</originalsourceid><addsrcrecordid>eNplUV2L1DAUDaLgMu4P8K3gc9d8tE3jW1nHdWEWkXHBt3Cb3sxk6DQ1ScX996Y7IoLJQy4n5xxOcgh5y-iNEIq-N_48LwlDZJwxypl8Qa44laKsBPv-8p_5NbmO8UTzUky0nF2RZTsdYTJuOhTdkvwZEg7F3viwIt4WD5COxR5HW25_zSNMkJyfiq8LjC49FY9xpe12D-UdThiexR8hQcQUPxRdFp5duV9mDD9dzHfdPAcP5viGvLIwRrz-c27I46ftt9vP5e7L3f1ttytNxVkq615wWUuhVNNUFdIBuFIKBUVkPSprwUiOZugHTttGUjswUUlVN6I3tGdKbMj9xXfwcNJzcGcIT9qD08-ADwcNITkzokao-57KxjacVjXIFhurYDUR0LJmyF7vLl75CT8WjEmf_BKmHF_zVgkqqGrbzLq5sA6QTd1kfQpg8h7yVxg_oXUZ76TMdbRr0g1hF4EJPsaA9m9MRvXarv6vXfEbOHqZwA</addsrcrecordid><sourcetype>Open Website</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2893030988</pqid></control><display><type>article</type><title>Enhancing Automated Scoring of Math Self-Explanation Quality Using LLM-Generated Datasets: A Semi-Supervised Approach</title><source>Publicly Available Content Database</source><creator>Nakamoto, Ryosuke ; Flanagan, Brendan ; Yamauchi, Taisei ; Dai, Yiling ; Takami, Kyosuke ; Ogata, Hiroaki</creator><creatorcontrib>Nakamoto, Ryosuke ; Flanagan, Brendan ; Yamauchi, Taisei ; Dai, Yiling ; Takami, Kyosuke ; Ogata, Hiroaki</creatorcontrib><description>In the realm of mathematics education, self-explanation stands as a crucial learning mechanism, allowing learners to articulate their comprehension of intricate mathematical concepts and strategies. As digital learning platforms grow in prominence, there are mounting opportunities to collect and utilize mathematical self-explanations. However, these opportunities are met with challenges in automated evaluation. Automatic scoring of mathematical self-explanations is crucial for preprocessing tasks, including the categorization of learner responses, identification of common misconceptions, and the creation of tailored feedback and model solutions. Nevertheless, this task is hindered by the dearth of ample sample sets. Our research introduces a semi-supervised technique using the large language model (LLM), specifically its Japanese variant, to enrich datasets for the automated scoring of mathematical self-explanations. We rigorously evaluated the quality of self-explanations across five datasets, ranging from human-evaluated originals to ones devoid of original content. Our results show that combining LLM-based explanations with mathematical material significantly improves the model’s accuracy. Interestingly, there is an optimal limit to how many synthetic self-explanation data can benefit the system. Exceeding this limit does not further improve outcomes. This study thus highlights the need for careful consideration when integrating synthetic data into solutions, especially within the mathematics discipline.</description><identifier>ISSN: 2073-431X</identifier><identifier>EISSN: 2073-431X</identifier><identifier>DOI: 10.3390/computers12110217</identifier><language>eng</language><publisher>Basel: MDPI AG</publisher><subject>Accuracy ; Analysis ; automated scoring ; Automation ; Computational linguistics ; data augmentation ; Datasets ; Education ; Language ; language learning model (LLM) ; Language processing ; Large language models ; Machine learning ; Mechanization ; Model accuracy ; Natural language interfaces ; Neural networks ; Problem solving ; self-explanation ; semi-supervised learning ; Students ; Synthetic data</subject><ispartof>Computers (Basel), 2023-11, Vol.12 (11), p.217</ispartof><rights>COPYRIGHT 2023 MDPI AG</rights><rights>2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c421t-5b327573996644e0da2999e30ee1be9ffac72ecdbd208670fd13479563bc0b193</citedby><cites>FETCH-LOGICAL-c421t-5b327573996644e0da2999e30ee1be9ffac72ecdbd208670fd13479563bc0b193</cites><orcidid>0000-0001-9900-8763 ; 0000-0001-7644-997X ; 0000-0002-3384-7916</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://www.proquest.com/docview/2893030988/fulltextPDF?pq-origsite=primo$$EPDF$$P50$$Gproquest$$Hfree_for_read</linktopdf><linktohtml>$$Uhttps://www.proquest.com/docview/2893030988?pq-origsite=primo$$EHTML$$P50$$Gproquest$$Hfree_for_read</linktohtml><link.rule.ids>314,780,784,25753,27924,27925,37012,44590,75126</link.rule.ids></links><search><creatorcontrib>Nakamoto, Ryosuke</creatorcontrib><creatorcontrib>Flanagan, Brendan</creatorcontrib><creatorcontrib>Yamauchi, Taisei</creatorcontrib><creatorcontrib>Dai, Yiling</creatorcontrib><creatorcontrib>Takami, Kyosuke</creatorcontrib><creatorcontrib>Ogata, Hiroaki</creatorcontrib><title>Enhancing Automated Scoring of Math Self-Explanation Quality Using LLM-Generated Datasets: A Semi-Supervised Approach</title><title>Computers (Basel)</title><description>In the realm of mathematics education, self-explanation stands as a crucial learning mechanism, allowing learners to articulate their comprehension of intricate mathematical concepts and strategies. As digital learning platforms grow in prominence, there are mounting opportunities to collect and utilize mathematical self-explanations. However, these opportunities are met with challenges in automated evaluation. Automatic scoring of mathematical self-explanations is crucial for preprocessing tasks, including the categorization of learner responses, identification of common misconceptions, and the creation of tailored feedback and model solutions. Nevertheless, this task is hindered by the dearth of ample sample sets. Our research introduces a semi-supervised technique using the large language model (LLM), specifically its Japanese variant, to enrich datasets for the automated scoring of mathematical self-explanations. We rigorously evaluated the quality of self-explanations across five datasets, ranging from human-evaluated originals to ones devoid of original content. Our results show that combining LLM-based explanations with mathematical material significantly improves the model’s accuracy. Interestingly, there is an optimal limit to how many synthetic self-explanation data can benefit the system. Exceeding this limit does not further improve outcomes. This study thus highlights the need for careful consideration when integrating synthetic data into solutions, especially within the mathematics discipline.</description><subject>Accuracy</subject><subject>Analysis</subject><subject>automated scoring</subject><subject>Automation</subject><subject>Computational linguistics</subject><subject>data augmentation</subject><subject>Datasets</subject><subject>Education</subject><subject>Language</subject><subject>language learning model (LLM)</subject><subject>Language processing</subject><subject>Large language models</subject><subject>Machine learning</subject><subject>Mechanization</subject><subject>Model accuracy</subject><subject>Natural language interfaces</subject><subject>Neural networks</subject><subject>Problem solving</subject><subject>self-explanation</subject><subject>semi-supervised learning</subject><subject>Students</subject><subject>Synthetic data</subject><issn>2073-431X</issn><issn>2073-431X</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>PIMPY</sourceid><sourceid>DOA</sourceid><recordid>eNplUV2L1DAUDaLgMu4P8K3gc9d8tE3jW1nHdWEWkXHBt3Cb3sxk6DQ1ScX996Y7IoLJQy4n5xxOcgh5y-iNEIq-N_48LwlDZJwxypl8Qa44laKsBPv-8p_5NbmO8UTzUky0nF2RZTsdYTJuOhTdkvwZEg7F3viwIt4WD5COxR5HW25_zSNMkJyfiq8LjC49FY9xpe12D-UdThiexR8hQcQUPxRdFp5duV9mDD9dzHfdPAcP5viGvLIwRrz-c27I46ftt9vP5e7L3f1ttytNxVkq615wWUuhVNNUFdIBuFIKBUVkPSprwUiOZugHTttGUjswUUlVN6I3tGdKbMj9xXfwcNJzcGcIT9qD08-ADwcNITkzokao-57KxjacVjXIFhurYDUR0LJmyF7vLl75CT8WjEmf_BKmHF_zVgkqqGrbzLq5sA6QTd1kfQpg8h7yVxg_oXUZ76TMdbRr0g1hF4EJPsaA9m9MRvXarv6vXfEbOHqZwA</recordid><startdate>20231101</startdate><enddate>20231101</enddate><creator>Nakamoto, Ryosuke</creator><creator>Flanagan, Brendan</creator><creator>Yamauchi, Taisei</creator><creator>Dai, Yiling</creator><creator>Takami, Kyosuke</creator><creator>Ogata, Hiroaki</creator><general>MDPI AG</general><scope>AAYXX</scope><scope>CITATION</scope><scope>3V.</scope><scope>7SC</scope><scope>7XB</scope><scope>8AL</scope><scope>8FD</scope><scope>8FE</scope><scope>8FG</scope><scope>8FK</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>ARAPS</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>GNUQQ</scope><scope>HCIFZ</scope><scope>JQ2</scope><scope>K7-</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>M0N</scope><scope>P5Z</scope><scope>P62</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>Q9U</scope><scope>DOA</scope><orcidid>https://orcid.org/0000-0001-9900-8763</orcidid><orcidid>https://orcid.org/0000-0001-7644-997X</orcidid><orcidid>https://orcid.org/0000-0002-3384-7916</orcidid></search><sort><creationdate>20231101</creationdate><title>Enhancing Automated Scoring of Math Self-Explanation Quality Using LLM-Generated Datasets: A Semi-Supervised Approach</title><author>Nakamoto, Ryosuke ; Flanagan, Brendan ; Yamauchi, Taisei ; Dai, Yiling ; Takami, Kyosuke ; Ogata, Hiroaki</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c421t-5b327573996644e0da2999e30ee1be9ffac72ecdbd208670fd13479563bc0b193</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Accuracy</topic><topic>Analysis</topic><topic>automated scoring</topic><topic>Automation</topic><topic>Computational linguistics</topic><topic>data augmentation</topic><topic>Datasets</topic><topic>Education</topic><topic>Language</topic><topic>language learning model (LLM)</topic><topic>Language processing</topic><topic>Large language models</topic><topic>Machine learning</topic><topic>Mechanization</topic><topic>Model accuracy</topic><topic>Natural language interfaces</topic><topic>Neural networks</topic><topic>Problem solving</topic><topic>self-explanation</topic><topic>semi-supervised learning</topic><topic>Students</topic><topic>Synthetic data</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Nakamoto, Ryosuke</creatorcontrib><creatorcontrib>Flanagan, Brendan</creatorcontrib><creatorcontrib>Yamauchi, Taisei</creatorcontrib><creatorcontrib>Dai, Yiling</creatorcontrib><creatorcontrib>Takami, Kyosuke</creatorcontrib><creatorcontrib>Ogata, Hiroaki</creatorcontrib><collection>CrossRef</collection><collection>ProQuest Central (Corporate)</collection><collection>Computer and Information Systems Abstracts</collection><collection>ProQuest Central (purchase pre-March 2016)</collection><collection>Computing Database (Alumni Edition)</collection><collection>Technology Research Database</collection><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>ProQuest Central (Alumni) (purchase pre-March 2016)</collection><collection>ProQuest Central (Alumni)</collection><collection>ProQuest Central</collection><collection>Advanced Technologies &amp; Aerospace Collection</collection><collection>ProQuest Central Essentials</collection><collection>AUTh Library subscriptions: ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central</collection><collection>ProQuest Central Student</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Computer Science Collection</collection><collection>Computer Science Database</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>Computing Database</collection><collection>Advanced Technologies &amp; Aerospace Database</collection><collection>ProQuest Advanced Technologies &amp; Aerospace Collection</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>ProQuest Central Basic</collection><collection>DOAJ Directory of Open Access Journals</collection><jtitle>Computers (Basel)</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Nakamoto, Ryosuke</au><au>Flanagan, Brendan</au><au>Yamauchi, Taisei</au><au>Dai, Yiling</au><au>Takami, Kyosuke</au><au>Ogata, Hiroaki</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Enhancing Automated Scoring of Math Self-Explanation Quality Using LLM-Generated Datasets: A Semi-Supervised Approach</atitle><jtitle>Computers (Basel)</jtitle><date>2023-11-01</date><risdate>2023</risdate><volume>12</volume><issue>11</issue><spage>217</spage><pages>217-</pages><issn>2073-431X</issn><eissn>2073-431X</eissn><abstract>In the realm of mathematics education, self-explanation stands as a crucial learning mechanism, allowing learners to articulate their comprehension of intricate mathematical concepts and strategies. As digital learning platforms grow in prominence, there are mounting opportunities to collect and utilize mathematical self-explanations. However, these opportunities are met with challenges in automated evaluation. Automatic scoring of mathematical self-explanations is crucial for preprocessing tasks, including the categorization of learner responses, identification of common misconceptions, and the creation of tailored feedback and model solutions. Nevertheless, this task is hindered by the dearth of ample sample sets. Our research introduces a semi-supervised technique using the large language model (LLM), specifically its Japanese variant, to enrich datasets for the automated scoring of mathematical self-explanations. We rigorously evaluated the quality of self-explanations across five datasets, ranging from human-evaluated originals to ones devoid of original content. Our results show that combining LLM-based explanations with mathematical material significantly improves the model’s accuracy. Interestingly, there is an optimal limit to how many synthetic self-explanation data can benefit the system. Exceeding this limit does not further improve outcomes. This study thus highlights the need for careful consideration when integrating synthetic data into solutions, especially within the mathematics discipline.</abstract><cop>Basel</cop><pub>MDPI AG</pub><doi>10.3390/computers12110217</doi><orcidid>https://orcid.org/0000-0001-9900-8763</orcidid><orcidid>https://orcid.org/0000-0001-7644-997X</orcidid><orcidid>https://orcid.org/0000-0002-3384-7916</orcidid><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 2073-431X
ispartof Computers (Basel), 2023-11, Vol.12 (11), p.217
issn 2073-431X
2073-431X
language eng
recordid cdi_doaj_primary_oai_doaj_org_article_ea5bb076f62045a78e6f9ab1933a816d
source Publicly Available Content Database
subjects Accuracy
Analysis
automated scoring
Automation
Computational linguistics
data augmentation
Datasets
Education
Language
language learning model (LLM)
Language processing
Large language models
Machine learning
Mechanization
Model accuracy
Natural language interfaces
Neural networks
Problem solving
self-explanation
semi-supervised learning
Students
Synthetic data
title Enhancing Automated Scoring of Math Self-Explanation Quality Using LLM-Generated Datasets: A Semi-Supervised Approach
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-04T21%3A39%3A01IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-gale_doaj_&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Enhancing%20Automated%20Scoring%20of%20Math%20Self-Explanation%20Quality%20Using%20LLM-Generated%20Datasets:%20A%20Semi-Supervised%20Approach&rft.jtitle=Computers%20(Basel)&rft.au=Nakamoto,%20Ryosuke&rft.date=2023-11-01&rft.volume=12&rft.issue=11&rft.spage=217&rft.pages=217-&rft.issn=2073-431X&rft.eissn=2073-431X&rft_id=info:doi/10.3390/computers12110217&rft_dat=%3Cgale_doaj_%3EA774318956%3C/gale_doaj_%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c421t-5b327573996644e0da2999e30ee1be9ffac72ecdbd208670fd13479563bc0b193%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2893030988&rft_id=info:pmid/&rft_galeid=A774318956&rfr_iscdi=true