Loading…
DCR: Divide-and-Conquer Reasoning for Multi-choice Question Answering with LLMs
Large language models (LLMs) have shown impressive performance in reasoning benchmarks with the emergence of Chain-of-Thought (CoT), particularly in multi-choice question (MCQ). However, current works equally resolve questions regardless of the problem-solving difficulty, leading to an excessive foc...
Saved in:
Published in: | arXiv.org 2024-04 |
---|---|
Main Authors: | , , , |
Format: | Article |
Language: | English |
Subjects: | |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
cited_by | |
---|---|
cites | |
container_end_page | |
container_issue | |
container_start_page | |
container_title | arXiv.org |
container_volume | |
creator | Meng, Zijie Zhang, Yan Feng, Zhaopeng Liu, Zuozhu |
description | Large language models (LLMs) have shown impressive performance in reasoning benchmarks with the emergence of Chain-of-Thought (CoT), particularly in multi-choice question (MCQ). However, current works equally resolve questions regardless of the problem-solving difficulty, leading to an excessive focus on simple items while insufficient attention on intricate ones. To address this challenge, we propose a simple yet effective strategy, Divide and Conquer Reasoning (DCR), to enhance the reasoning capability of LLMs for MCQs, as inspired by human beings using heuristics to first categorize tasks and then handle them separately. In particular, we first categorize questions into two subsets based on confidence score (\(\mathcal{CS}\)), which is estimated by statistical frequency of generated answers. Subsequently, we propose Filter Choices based Reasoning (FCR) to improve model performance on MCQs with low (\(\mathcal{CS}\)). Our experiments demonstrate that the proposed strategy only costs 85% of SOTA, while still achieves average accuracy improvement of 1.56% across nine datasets including arithmetic, commonsense, and logic reasoning tasks. The code is at \url{https://github.com/AiMijie/Divide-and-Conquer} |
format | article |
fullrecord | <record><control><sourceid>proquest</sourceid><recordid>TN_cdi_proquest_journals_2913255411</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2913255411</sourcerecordid><originalsourceid>FETCH-proquest_journals_29132554113</originalsourceid><addsrcrecordid>eNqNjt8KgjAcRkcQJOU7DLoeuE37dxdadKFE0r0MnTmR33Jz-fop9ABdnYvvfHAWyGOcU3IIGVsh39o2CAK227Mo4h66J3F-won6qEoSARWJNfROGpxLYTUoeOFaG5y5blCkbLQqJX44aQelAZ_BjtLMzqiGBqdpZjdoWYvOSv_HNdpeL8_4Rt5G9_OvaLUzME0FO1I-RYSU8v-sL4kCPZI</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2913255411</pqid></control><display><type>article</type><title>DCR: Divide-and-Conquer Reasoning for Multi-choice Question Answering with LLMs</title><source>Publicly Available Content Database</source><creator>Meng, Zijie ; Zhang, Yan ; Feng, Zhaopeng ; Liu, Zuozhu</creator><creatorcontrib>Meng, Zijie ; Zhang, Yan ; Feng, Zhaopeng ; Liu, Zuozhu</creatorcontrib><description>Large language models (LLMs) have shown impressive performance in reasoning benchmarks with the emergence of Chain-of-Thought (CoT), particularly in multi-choice question (MCQ). However, current works equally resolve questions regardless of the problem-solving difficulty, leading to an excessive focus on simple items while insufficient attention on intricate ones. To address this challenge, we propose a simple yet effective strategy, Divide and Conquer Reasoning (DCR), to enhance the reasoning capability of LLMs for MCQs, as inspired by human beings using heuristics to first categorize tasks and then handle them separately. In particular, we first categorize questions into two subsets based on confidence score (\(\mathcal{CS}\)), which is estimated by statistical frequency of generated answers. Subsequently, we propose Filter Choices based Reasoning (FCR) to improve model performance on MCQs with low (\(\mathcal{CS}\)). Our experiments demonstrate that the proposed strategy only costs 85% of SOTA, while still achieves average accuracy improvement of 1.56% across nine datasets including arithmetic, commonsense, and logic reasoning tasks. The code is at \url{https://github.com/AiMijie/Divide-and-Conquer}</description><identifier>EISSN: 2331-8422</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Confidence ; Large language models ; Questions ; Reasoning</subject><ispartof>arXiv.org, 2024-04</ispartof><rights>2024. This work is published under http://arxiv.org/licenses/nonexclusive-distrib/1.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://www.proquest.com/docview/2913255411?pq-origsite=primo$$EHTML$$P50$$Gproquest$$Hfree_for_read</linktohtml><link.rule.ids>780,784,25751,37010,44588</link.rule.ids></links><search><creatorcontrib>Meng, Zijie</creatorcontrib><creatorcontrib>Zhang, Yan</creatorcontrib><creatorcontrib>Feng, Zhaopeng</creatorcontrib><creatorcontrib>Liu, Zuozhu</creatorcontrib><title>DCR: Divide-and-Conquer Reasoning for Multi-choice Question Answering with LLMs</title><title>arXiv.org</title><description>Large language models (LLMs) have shown impressive performance in reasoning benchmarks with the emergence of Chain-of-Thought (CoT), particularly in multi-choice question (MCQ). However, current works equally resolve questions regardless of the problem-solving difficulty, leading to an excessive focus on simple items while insufficient attention on intricate ones. To address this challenge, we propose a simple yet effective strategy, Divide and Conquer Reasoning (DCR), to enhance the reasoning capability of LLMs for MCQs, as inspired by human beings using heuristics to first categorize tasks and then handle them separately. In particular, we first categorize questions into two subsets based on confidence score (\(\mathcal{CS}\)), which is estimated by statistical frequency of generated answers. Subsequently, we propose Filter Choices based Reasoning (FCR) to improve model performance on MCQs with low (\(\mathcal{CS}\)). Our experiments demonstrate that the proposed strategy only costs 85% of SOTA, while still achieves average accuracy improvement of 1.56% across nine datasets including arithmetic, commonsense, and logic reasoning tasks. The code is at \url{https://github.com/AiMijie/Divide-and-Conquer}</description><subject>Confidence</subject><subject>Large language models</subject><subject>Questions</subject><subject>Reasoning</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>PIMPY</sourceid><recordid>eNqNjt8KgjAcRkcQJOU7DLoeuE37dxdadKFE0r0MnTmR33Jz-fop9ABdnYvvfHAWyGOcU3IIGVsh39o2CAK227Mo4h66J3F-won6qEoSARWJNfROGpxLYTUoeOFaG5y5blCkbLQqJX44aQelAZ_BjtLMzqiGBqdpZjdoWYvOSv_HNdpeL8_4Rt5G9_OvaLUzME0FO1I-RYSU8v-sL4kCPZI</recordid><startdate>20240402</startdate><enddate>20240402</enddate><creator>Meng, Zijie</creator><creator>Zhang, Yan</creator><creator>Feng, Zhaopeng</creator><creator>Liu, Zuozhu</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope></search><sort><creationdate>20240402</creationdate><title>DCR: Divide-and-Conquer Reasoning for Multi-choice Question Answering with LLMs</title><author>Meng, Zijie ; Zhang, Yan ; Feng, Zhaopeng ; Liu, Zuozhu</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-proquest_journals_29132554113</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Confidence</topic><topic>Large language models</topic><topic>Questions</topic><topic>Reasoning</topic><toplevel>online_resources</toplevel><creatorcontrib>Meng, Zijie</creatorcontrib><creatorcontrib>Zhang, Yan</creatorcontrib><creatorcontrib>Feng, Zhaopeng</creatorcontrib><creatorcontrib>Liu, Zuozhu</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science & Engineering Collection</collection><collection>ProQuest Central (Alumni)</collection><collection>ProQuest Central</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering Collection</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Meng, Zijie</au><au>Zhang, Yan</au><au>Feng, Zhaopeng</au><au>Liu, Zuozhu</au><format>book</format><genre>document</genre><ristype>GEN</ristype><atitle>DCR: Divide-and-Conquer Reasoning for Multi-choice Question Answering with LLMs</atitle><jtitle>arXiv.org</jtitle><date>2024-04-02</date><risdate>2024</risdate><eissn>2331-8422</eissn><abstract>Large language models (LLMs) have shown impressive performance in reasoning benchmarks with the emergence of Chain-of-Thought (CoT), particularly in multi-choice question (MCQ). However, current works equally resolve questions regardless of the problem-solving difficulty, leading to an excessive focus on simple items while insufficient attention on intricate ones. To address this challenge, we propose a simple yet effective strategy, Divide and Conquer Reasoning (DCR), to enhance the reasoning capability of LLMs for MCQs, as inspired by human beings using heuristics to first categorize tasks and then handle them separately. In particular, we first categorize questions into two subsets based on confidence score (\(\mathcal{CS}\)), which is estimated by statistical frequency of generated answers. Subsequently, we propose Filter Choices based Reasoning (FCR) to improve model performance on MCQs with low (\(\mathcal{CS}\)). Our experiments demonstrate that the proposed strategy only costs 85% of SOTA, while still achieves average accuracy improvement of 1.56% across nine datasets including arithmetic, commonsense, and logic reasoning tasks. The code is at \url{https://github.com/AiMijie/Divide-and-Conquer}</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | EISSN: 2331-8422 |
ispartof | arXiv.org, 2024-04 |
issn | 2331-8422 |
language | eng |
recordid | cdi_proquest_journals_2913255411 |
source | Publicly Available Content Database |
subjects | Confidence Large language models Questions Reasoning |
title | DCR: Divide-and-Conquer Reasoning for Multi-choice Question Answering with LLMs |
url | http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-10T06%3A45%3A22IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=document&rft.atitle=DCR:%20Divide-and-Conquer%20Reasoning%20for%20Multi-choice%20Question%20Answering%20with%20LLMs&rft.jtitle=arXiv.org&rft.au=Meng,%20Zijie&rft.date=2024-04-02&rft.eissn=2331-8422&rft_id=info:doi/&rft_dat=%3Cproquest%3E2913255411%3C/proquest%3E%3Cgrp_id%3Ecdi_FETCH-proquest_journals_29132554113%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2913255411&rft_id=info:pmid/&rfr_iscdi=true |