Loading…
Enhancing Automated Scoring of Math Self-Explanation Quality Using LLM-Generated Datasets: A Semi-Supervised Approach
In the realm of mathematics education, self-explanation stands as a crucial learning mechanism, allowing learners to articulate their comprehension of intricate mathematical concepts and strategies. As digital learning platforms grow in prominence, there are mounting opportunities to collect and uti...
Saved in:
Published in: | Computers (Basel) 2023-11, Vol.12 (11), p.217 |
---|---|
Main Authors: | , , , , , |
Format: | Article |
Language: | English |
Subjects: | |
Citations: | Items that this one cites Items that cite this one |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
cited_by | cdi_FETCH-LOGICAL-c421t-5b327573996644e0da2999e30ee1be9ffac72ecdbd208670fd13479563bc0b193 |
---|---|
cites | cdi_FETCH-LOGICAL-c421t-5b327573996644e0da2999e30ee1be9ffac72ecdbd208670fd13479563bc0b193 |
container_end_page | |
container_issue | 11 |
container_start_page | 217 |
container_title | Computers (Basel) |
container_volume | 12 |
creator | Nakamoto, Ryosuke Flanagan, Brendan Yamauchi, Taisei Dai, Yiling Takami, Kyosuke Ogata, Hiroaki |
description | In the realm of mathematics education, self-explanation stands as a crucial learning mechanism, allowing learners to articulate their comprehension of intricate mathematical concepts and strategies. As digital learning platforms grow in prominence, there are mounting opportunities to collect and utilize mathematical self-explanations. However, these opportunities are met with challenges in automated evaluation. Automatic scoring of mathematical self-explanations is crucial for preprocessing tasks, including the categorization of learner responses, identification of common misconceptions, and the creation of tailored feedback and model solutions. Nevertheless, this task is hindered by the dearth of ample sample sets. Our research introduces a semi-supervised technique using the large language model (LLM), specifically its Japanese variant, to enrich datasets for the automated scoring of mathematical self-explanations. We rigorously evaluated the quality of self-explanations across five datasets, ranging from human-evaluated originals to ones devoid of original content. Our results show that combining LLM-based explanations with mathematical material significantly improves the model’s accuracy. Interestingly, there is an optimal limit to how many synthetic self-explanation data can benefit the system. Exceeding this limit does not further improve outcomes. This study thus highlights the need for careful consideration when integrating synthetic data into solutions, especially within the mathematics discipline. |
doi_str_mv | 10.3390/computers12110217 |
format | article |
fullrecord | <record><control><sourceid>gale_doaj_</sourceid><recordid>TN_cdi_doaj_primary_oai_doaj_org_article_ea5bb076f62045a78e6f9ab1933a816d</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><galeid>A774318956</galeid><doaj_id>oai_doaj_org_article_ea5bb076f62045a78e6f9ab1933a816d</doaj_id><sourcerecordid>A774318956</sourcerecordid><originalsourceid>FETCH-LOGICAL-c421t-5b327573996644e0da2999e30ee1be9ffac72ecdbd208670fd13479563bc0b193</originalsourceid><addsrcrecordid>eNplUV2L1DAUDaLgMu4P8K3gc9d8tE3jW1nHdWEWkXHBt3Cb3sxk6DQ1ScX996Y7IoLJQy4n5xxOcgh5y-iNEIq-N_48LwlDZJwxypl8Qa44laKsBPv-8p_5NbmO8UTzUky0nF2RZTsdYTJuOhTdkvwZEg7F3viwIt4WD5COxR5HW25_zSNMkJyfiq8LjC49FY9xpe12D-UdThiexR8hQcQUPxRdFp5duV9mDD9dzHfdPAcP5viGvLIwRrz-c27I46ftt9vP5e7L3f1ttytNxVkq615wWUuhVNNUFdIBuFIKBUVkPSprwUiOZugHTttGUjswUUlVN6I3tGdKbMj9xXfwcNJzcGcIT9qD08-ADwcNITkzokao-57KxjacVjXIFhurYDUR0LJmyF7vLl75CT8WjEmf_BKmHF_zVgkqqGrbzLq5sA6QTd1kfQpg8h7yVxg_oXUZ76TMdbRr0g1hF4EJPsaA9m9MRvXarv6vXfEbOHqZwA</addsrcrecordid><sourcetype>Open Website</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2893030988</pqid></control><display><type>article</type><title>Enhancing Automated Scoring of Math Self-Explanation Quality Using LLM-Generated Datasets: A Semi-Supervised Approach</title><source>Publicly Available Content Database</source><creator>Nakamoto, Ryosuke ; Flanagan, Brendan ; Yamauchi, Taisei ; Dai, Yiling ; Takami, Kyosuke ; Ogata, Hiroaki</creator><creatorcontrib>Nakamoto, Ryosuke ; Flanagan, Brendan ; Yamauchi, Taisei ; Dai, Yiling ; Takami, Kyosuke ; Ogata, Hiroaki</creatorcontrib><description>In the realm of mathematics education, self-explanation stands as a crucial learning mechanism, allowing learners to articulate their comprehension of intricate mathematical concepts and strategies. As digital learning platforms grow in prominence, there are mounting opportunities to collect and utilize mathematical self-explanations. However, these opportunities are met with challenges in automated evaluation. Automatic scoring of mathematical self-explanations is crucial for preprocessing tasks, including the categorization of learner responses, identification of common misconceptions, and the creation of tailored feedback and model solutions. Nevertheless, this task is hindered by the dearth of ample sample sets. Our research introduces a semi-supervised technique using the large language model (LLM), specifically its Japanese variant, to enrich datasets for the automated scoring of mathematical self-explanations. We rigorously evaluated the quality of self-explanations across five datasets, ranging from human-evaluated originals to ones devoid of original content. Our results show that combining LLM-based explanations with mathematical material significantly improves the model’s accuracy. Interestingly, there is an optimal limit to how many synthetic self-explanation data can benefit the system. Exceeding this limit does not further improve outcomes. This study thus highlights the need for careful consideration when integrating synthetic data into solutions, especially within the mathematics discipline.</description><identifier>ISSN: 2073-431X</identifier><identifier>EISSN: 2073-431X</identifier><identifier>DOI: 10.3390/computers12110217</identifier><language>eng</language><publisher>Basel: MDPI AG</publisher><subject>Accuracy ; Analysis ; automated scoring ; Automation ; Computational linguistics ; data augmentation ; Datasets ; Education ; Language ; language learning model (LLM) ; Language processing ; Large language models ; Machine learning ; Mechanization ; Model accuracy ; Natural language interfaces ; Neural networks ; Problem solving ; self-explanation ; semi-supervised learning ; Students ; Synthetic data</subject><ispartof>Computers (Basel), 2023-11, Vol.12 (11), p.217</ispartof><rights>COPYRIGHT 2023 MDPI AG</rights><rights>2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c421t-5b327573996644e0da2999e30ee1be9ffac72ecdbd208670fd13479563bc0b193</citedby><cites>FETCH-LOGICAL-c421t-5b327573996644e0da2999e30ee1be9ffac72ecdbd208670fd13479563bc0b193</cites><orcidid>0000-0001-9900-8763 ; 0000-0001-7644-997X ; 0000-0002-3384-7916</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://www.proquest.com/docview/2893030988/fulltextPDF?pq-origsite=primo$$EPDF$$P50$$Gproquest$$Hfree_for_read</linktopdf><linktohtml>$$Uhttps://www.proquest.com/docview/2893030988?pq-origsite=primo$$EHTML$$P50$$Gproquest$$Hfree_for_read</linktohtml><link.rule.ids>314,780,784,25753,27924,27925,37012,44590,75126</link.rule.ids></links><search><creatorcontrib>Nakamoto, Ryosuke</creatorcontrib><creatorcontrib>Flanagan, Brendan</creatorcontrib><creatorcontrib>Yamauchi, Taisei</creatorcontrib><creatorcontrib>Dai, Yiling</creatorcontrib><creatorcontrib>Takami, Kyosuke</creatorcontrib><creatorcontrib>Ogata, Hiroaki</creatorcontrib><title>Enhancing Automated Scoring of Math Self-Explanation Quality Using LLM-Generated Datasets: A Semi-Supervised Approach</title><title>Computers (Basel)</title><description>In the realm of mathematics education, self-explanation stands as a crucial learning mechanism, allowing learners to articulate their comprehension of intricate mathematical concepts and strategies. As digital learning platforms grow in prominence, there are mounting opportunities to collect and utilize mathematical self-explanations. However, these opportunities are met with challenges in automated evaluation. Automatic scoring of mathematical self-explanations is crucial for preprocessing tasks, including the categorization of learner responses, identification of common misconceptions, and the creation of tailored feedback and model solutions. Nevertheless, this task is hindered by the dearth of ample sample sets. Our research introduces a semi-supervised technique using the large language model (LLM), specifically its Japanese variant, to enrich datasets for the automated scoring of mathematical self-explanations. We rigorously evaluated the quality of self-explanations across five datasets, ranging from human-evaluated originals to ones devoid of original content. Our results show that combining LLM-based explanations with mathematical material significantly improves the model’s accuracy. Interestingly, there is an optimal limit to how many synthetic self-explanation data can benefit the system. Exceeding this limit does not further improve outcomes. This study thus highlights the need for careful consideration when integrating synthetic data into solutions, especially within the mathematics discipline.</description><subject>Accuracy</subject><subject>Analysis</subject><subject>automated scoring</subject><subject>Automation</subject><subject>Computational linguistics</subject><subject>data augmentation</subject><subject>Datasets</subject><subject>Education</subject><subject>Language</subject><subject>language learning model (LLM)</subject><subject>Language processing</subject><subject>Large language models</subject><subject>Machine learning</subject><subject>Mechanization</subject><subject>Model accuracy</subject><subject>Natural language interfaces</subject><subject>Neural networks</subject><subject>Problem solving</subject><subject>self-explanation</subject><subject>semi-supervised learning</subject><subject>Students</subject><subject>Synthetic data</subject><issn>2073-431X</issn><issn>2073-431X</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>PIMPY</sourceid><sourceid>DOA</sourceid><recordid>eNplUV2L1DAUDaLgMu4P8K3gc9d8tE3jW1nHdWEWkXHBt3Cb3sxk6DQ1ScX996Y7IoLJQy4n5xxOcgh5y-iNEIq-N_48LwlDZJwxypl8Qa44laKsBPv-8p_5NbmO8UTzUky0nF2RZTsdYTJuOhTdkvwZEg7F3viwIt4WD5COxR5HW25_zSNMkJyfiq8LjC49FY9xpe12D-UdThiexR8hQcQUPxRdFp5duV9mDD9dzHfdPAcP5viGvLIwRrz-c27I46ftt9vP5e7L3f1ttytNxVkq615wWUuhVNNUFdIBuFIKBUVkPSprwUiOZugHTttGUjswUUlVN6I3tGdKbMj9xXfwcNJzcGcIT9qD08-ADwcNITkzokao-57KxjacVjXIFhurYDUR0LJmyF7vLl75CT8WjEmf_BKmHF_zVgkqqGrbzLq5sA6QTd1kfQpg8h7yVxg_oXUZ76TMdbRr0g1hF4EJPsaA9m9MRvXarv6vXfEbOHqZwA</recordid><startdate>20231101</startdate><enddate>20231101</enddate><creator>Nakamoto, Ryosuke</creator><creator>Flanagan, Brendan</creator><creator>Yamauchi, Taisei</creator><creator>Dai, Yiling</creator><creator>Takami, Kyosuke</creator><creator>Ogata, Hiroaki</creator><general>MDPI AG</general><scope>AAYXX</scope><scope>CITATION</scope><scope>3V.</scope><scope>7SC</scope><scope>7XB</scope><scope>8AL</scope><scope>8FD</scope><scope>8FE</scope><scope>8FG</scope><scope>8FK</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>ARAPS</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>GNUQQ</scope><scope>HCIFZ</scope><scope>JQ2</scope><scope>K7-</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>M0N</scope><scope>P5Z</scope><scope>P62</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>Q9U</scope><scope>DOA</scope><orcidid>https://orcid.org/0000-0001-9900-8763</orcidid><orcidid>https://orcid.org/0000-0001-7644-997X</orcidid><orcidid>https://orcid.org/0000-0002-3384-7916</orcidid></search><sort><creationdate>20231101</creationdate><title>Enhancing Automated Scoring of Math Self-Explanation Quality Using LLM-Generated Datasets: A Semi-Supervised Approach</title><author>Nakamoto, Ryosuke ; Flanagan, Brendan ; Yamauchi, Taisei ; Dai, Yiling ; Takami, Kyosuke ; Ogata, Hiroaki</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c421t-5b327573996644e0da2999e30ee1be9ffac72ecdbd208670fd13479563bc0b193</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Accuracy</topic><topic>Analysis</topic><topic>automated scoring</topic><topic>Automation</topic><topic>Computational linguistics</topic><topic>data augmentation</topic><topic>Datasets</topic><topic>Education</topic><topic>Language</topic><topic>language learning model (LLM)</topic><topic>Language processing</topic><topic>Large language models</topic><topic>Machine learning</topic><topic>Mechanization</topic><topic>Model accuracy</topic><topic>Natural language interfaces</topic><topic>Neural networks</topic><topic>Problem solving</topic><topic>self-explanation</topic><topic>semi-supervised learning</topic><topic>Students</topic><topic>Synthetic data</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Nakamoto, Ryosuke</creatorcontrib><creatorcontrib>Flanagan, Brendan</creatorcontrib><creatorcontrib>Yamauchi, Taisei</creatorcontrib><creatorcontrib>Dai, Yiling</creatorcontrib><creatorcontrib>Takami, Kyosuke</creatorcontrib><creatorcontrib>Ogata, Hiroaki</creatorcontrib><collection>CrossRef</collection><collection>ProQuest Central (Corporate)</collection><collection>Computer and Information Systems Abstracts</collection><collection>ProQuest Central (purchase pre-March 2016)</collection><collection>Computing Database (Alumni Edition)</collection><collection>Technology Research Database</collection><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>ProQuest Central (Alumni) (purchase pre-March 2016)</collection><collection>ProQuest Central (Alumni)</collection><collection>ProQuest Central</collection><collection>Advanced Technologies & Aerospace Collection</collection><collection>ProQuest Central Essentials</collection><collection>AUTh Library subscriptions: ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central</collection><collection>ProQuest Central Student</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Computer Science Collection</collection><collection>Computer Science Database</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>Computing Database</collection><collection>Advanced Technologies & Aerospace Database</collection><collection>ProQuest Advanced Technologies & Aerospace Collection</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>ProQuest Central Basic</collection><collection>DOAJ Directory of Open Access Journals</collection><jtitle>Computers (Basel)</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Nakamoto, Ryosuke</au><au>Flanagan, Brendan</au><au>Yamauchi, Taisei</au><au>Dai, Yiling</au><au>Takami, Kyosuke</au><au>Ogata, Hiroaki</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Enhancing Automated Scoring of Math Self-Explanation Quality Using LLM-Generated Datasets: A Semi-Supervised Approach</atitle><jtitle>Computers (Basel)</jtitle><date>2023-11-01</date><risdate>2023</risdate><volume>12</volume><issue>11</issue><spage>217</spage><pages>217-</pages><issn>2073-431X</issn><eissn>2073-431X</eissn><abstract>In the realm of mathematics education, self-explanation stands as a crucial learning mechanism, allowing learners to articulate their comprehension of intricate mathematical concepts and strategies. As digital learning platforms grow in prominence, there are mounting opportunities to collect and utilize mathematical self-explanations. However, these opportunities are met with challenges in automated evaluation. Automatic scoring of mathematical self-explanations is crucial for preprocessing tasks, including the categorization of learner responses, identification of common misconceptions, and the creation of tailored feedback and model solutions. Nevertheless, this task is hindered by the dearth of ample sample sets. Our research introduces a semi-supervised technique using the large language model (LLM), specifically its Japanese variant, to enrich datasets for the automated scoring of mathematical self-explanations. We rigorously evaluated the quality of self-explanations across five datasets, ranging from human-evaluated originals to ones devoid of original content. Our results show that combining LLM-based explanations with mathematical material significantly improves the model’s accuracy. Interestingly, there is an optimal limit to how many synthetic self-explanation data can benefit the system. Exceeding this limit does not further improve outcomes. This study thus highlights the need for careful consideration when integrating synthetic data into solutions, especially within the mathematics discipline.</abstract><cop>Basel</cop><pub>MDPI AG</pub><doi>10.3390/computers12110217</doi><orcidid>https://orcid.org/0000-0001-9900-8763</orcidid><orcidid>https://orcid.org/0000-0001-7644-997X</orcidid><orcidid>https://orcid.org/0000-0002-3384-7916</orcidid><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | ISSN: 2073-431X |
ispartof | Computers (Basel), 2023-11, Vol.12 (11), p.217 |
issn | 2073-431X 2073-431X |
language | eng |
recordid | cdi_doaj_primary_oai_doaj_org_article_ea5bb076f62045a78e6f9ab1933a816d |
source | Publicly Available Content Database |
subjects | Accuracy Analysis automated scoring Automation Computational linguistics data augmentation Datasets Education Language language learning model (LLM) Language processing Large language models Machine learning Mechanization Model accuracy Natural language interfaces Neural networks Problem solving self-explanation semi-supervised learning Students Synthetic data |
title | Enhancing Automated Scoring of Math Self-Explanation Quality Using LLM-Generated Datasets: A Semi-Supervised Approach |
url | http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-04T21%3A39%3A01IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-gale_doaj_&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Enhancing%20Automated%20Scoring%20of%20Math%20Self-Explanation%20Quality%20Using%20LLM-Generated%20Datasets:%20A%20Semi-Supervised%20Approach&rft.jtitle=Computers%20(Basel)&rft.au=Nakamoto,%20Ryosuke&rft.date=2023-11-01&rft.volume=12&rft.issue=11&rft.spage=217&rft.pages=217-&rft.issn=2073-431X&rft.eissn=2073-431X&rft_id=info:doi/10.3390/computers12110217&rft_dat=%3Cgale_doaj_%3EA774318956%3C/gale_doaj_%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c421t-5b327573996644e0da2999e30ee1be9ffac72ecdbd208670fd13479563bc0b193%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2893030988&rft_id=info:pmid/&rft_galeid=A774318956&rfr_iscdi=true |