Enhance Reasoning in Large Language Models via Problem Refinement (2024)

Zhiheng Xi11{}^{1}start_FLOATSUPERSCRIPT 1 end_FLOATSUPERSCRIPT, Senjie Jin1*1{}^{1*}start_FLOATSUPERSCRIPT 1 * end_FLOATSUPERSCRIPT, Yuhao Zhou11{}^{1}start_FLOATSUPERSCRIPT 1 end_FLOATSUPERSCRIPT, Rui Zheng11{}^{1}start_FLOATSUPERSCRIPT 1 end_FLOATSUPERSCRIPT, Songyang Gao11{}^{1}start_FLOATSUPERSCRIPT 1 end_FLOATSUPERSCRIPT,
Tao Gui22{}^{2}start_FLOATSUPERSCRIPT 2 end_FLOATSUPERSCRIPT , Qi Zhang11normal-†{}^{1{\dagger}}start_FLOATSUPERSCRIPT 1 † end_FLOATSUPERSCRIPT, Xuanjing Huang11{}^{1}start_FLOATSUPERSCRIPT 1 end_FLOATSUPERSCRIPT
11{}^{1}start_FLOATSUPERSCRIPT 1 end_FLOATSUPERSCRIPT School of Computer Science, Fudan University, Shanghai, China
22{}^{2}start_FLOATSUPERSCRIPT 2 end_FLOATSUPERSCRIPT Institute of Modern Languages and Linguistics, Fudan University, Shanghai, China
{zhxi22,sjjin22,zhouyh21,gaosy21}@m.fudan.edu.cn,
{rzheng20,tgui,qz,xjhuang}@fudan.edu.cn
Equal contribution. Corresponding author.

Abstract

To enhance the multi-step reasoning capabilities of large language models, researchers have extensively explored prompting methods, notably the Chain-of-Thought (CoT) method which explicitly elicits human-like rationales. However, they have inadvertently overlooked the potential of enhancing model reasoning performance by formulating higher-quality problems 111A reasoning problem often consists of two parts: the context and the final question Creswell etal. (2022).. In this work, we start from the problem side and propose Self-Polish (SP), a novel method that facilitates the model’s reasoning by guiding it to progressively refine the given problems to be more comprehensible and solvable. We also explore several automatic prompting varients and propose the Self-Polish prompt bank for the community. SP is orthogonal to all other prompting methods of answer/reasoning side like CoT, allowing for seamless integration with state-of-the-art techniques for further improvement. Thorough experiments show that the proposed method attains notable and consistent effectiveness on five reasoning benchmarks across different models. Furthermore, our method also showcases impressive performance on robustness evaluation. Codes and prompts are available at https://github.com/WooooDyy/Self-Polish.

1 Introduction

Enhance Reasoning in Large Language Models via Problem Refinement (1)

Large language models (LLMs) have achieved impressive performance on a variety of NLP tasks (Brown etal., 2020; Otter etal., 2021; Chowdhery etal., 2022), but their capability to perform multi-step reasoning is considered a limitation, which can not be tackled solely by scaling up the model size (Rae etal., 2021; Srivastava etal., 2022).To address this challenge, many prompting methods have been proposed to elicit reasoning in LLMs, and have demonstrated significant effectiveness (Wei etal., 2022b; Fu etal., 2022; Zhou etal., 2022a).

Chain-of-Thought (CoT) is a breakthrough method that teaches a language model to imitate the step-by-step reasoning process of humans to solve a reasoning task (Wei etal., 2022b).Many following work has explored variants of CoT to improve the quality of rationales of LLMs Kojima etal. (2022); Fu etal. (2022); Zhou etal. (2022a).There is also a line of work that optimizes the rationales for better consistency and continuity (Wang etal., 2022; Li etal., 2022; Zelikman etal., 2022; Zheng etal., 2023), and a representative one is Self-Consistency (SC). SC generates diverse reasoning paths and answers, and then leverages the majority vote strategy to get the most consistent answer (Wang etal., 2022).Despite the boosted reasoning performance of the aforementioned methods, they focus on the answer/reasoning side, and little emphasis has been placed on the problems side.

Actually, the clarity and logical structure of the problem description are crucial factors for human understanding and model comprehension (Shou and Smithson, 2015; Faruqui and Das, 2018; Chu etal., 2020).LLMs often exhibit poor reasoning performance when confronted with low-quality real-world reasoning problems, which may be excessively long, ambiguous, unclear in focus, or contain irrelevant information (Zellers etal., 2018; Shi etal., 2023; Ye and Durrett, 2022).To tackle this challenge, we consider refining problems into a better formulation.

In this work, we propose Self-Polish (Figure 1 right) that leverages LLMs themselves to refine reasoning problems without training for better reasoning performance.We first present several principles for refined problems: concise, clear, well-focused, and absent of irrelevant information.To achieve our goal, we propose the Self-Polish Prompt Bank which includes several feasible solutions as outlined in the following text.An intuitive strategy is to reformulate problems via instruction-following (Sanh etal., 2022; Ouyang etal., 2022),and we call it zero-shot problem refining.Next, we include demonstrations in the prompts (Brown etal., 2020; Chowdhery etal., 2022) to enable models to better internalize and apply the principles, which is defined as in-context problem refining.During the construction of the demonstrations, we incorporated a curated collection of problem-refining patterns, e.g., eliminating irrelevant information, rearranging the logic structure, and organizing local conditions into new ones in parallel.Moreover, we explore automatic prompting methods to construct enhanced prompts and mitigate manual efforts, based on the criteria of complexity (complexity-based Self-Polish) or diversity (automatic Self-Polish).To further enhance the reliability and consistency of the generated problems, we propose to progressively refine problems until obtaining a convergent answer.

Experiments show that our method consistently improves the reasoning performance of various models (i.e., Text-davinci-002, Text-davinci-003, GPT-3.5-Turbo) on five benchmarks (Table 1 & Figure 3).Moreover, the proposed method is orthogonal to all other reasoning-side state-of-the-art prompting methods, making it convenient to be combined with them for further improvement.Detailed experiments demonstrate that the performance of reasoning-side methods can be significantly boosted when integrated with SP (Table 2 & Table 3).Self-Polish also showcases exceptional performance on robustness evaluation (Figure 4).

In summary, we make the following contributions:

  1. 1.

    We propose a novel method, Self-Polish, to improve the reasoning performance and robustness of LLMs.

  2. 2.

    We demonstrate the effectiveness of our method when applied alone or combined with other prompting approaches on five benchmarks with different models.

  3. 3.

    We believe that the proposed Self-Polish represents an important step in enhancing LLMs’ reasoning capabilities by shifting the perspective from the answer/reasoning side to the problem side. We hope it could inspire future research in this field..

2 Related Work

Multi-step reasoning.

Multi-step reasoning tasks have posed significant challenges for language models (Rae etal., 2021; Bommasani etal., 2021; Qiao etal., 2022), and it is considered as an emergent ability of LLMs (Schaeffer etal., 2023).It is in these tasks that the effectiveness of few-shot prompting begins to surpass that of full training set fine-tuning (Lewkowycz etal., 2022).Moreover, such capability is considered important in building more complex artificial intelligence such as large language model-based agents (LLM-based agents) (Xi etal., 2023).Our work represents a significant stride in enhancing the ability of language models to perform multi-step reasoning tasks, through the facilitation of models’ comprehension and processing of given reasoning problems.

Reasoning with prompting.

Prompting strategies have substantially improved the reasoning ability of LLMs by a large margin (Qiao etal., 2022; Lewkowycz etal., 2022). An important line of work in this area is Chain-of-Thought (CoT) prompting which elicits the reasoning ability of models by prompting them to imitate the step-by-step reasoning process of humans (Wei etal., 2022b; Kojima etal., 2022; Fu etal., 2022; Zhou etal., 2022a).Another line of work focuses on optimizing the rationales for better consistency and continuity (Wang etal., 2022; Li etal., 2022; Zelikman etal., 2022; Zheng etal., 2023).A representative one is Self-Consistency (SC), which samples multiple reasoning paths and generate the most consistent answer by majority vote (Wang etal., 2022). Different from Self-Polish, the aforementioned strategies emphasize improving the quality of rationales from the answer/reasoning side.Our method is a problem-side method, so it is orthogonal to all of them and can be combined with them for further improvement.

See Appendix A for more related work and the detailed differences between Self-Polish and Least-to-Most Zhou etal. (2022a).

Enhance Reasoning in Large Language Models via Problem Refinement (2)

MethodProgressivelyDatasetAverage
GSM8KAQuASVAMPMultiArithMathQA
Standard15.815.815.815.828.328.328.328.372.972.972.972.935.135.135.135.128.228.228.228.236.136.136.136.1
+Zero-shot SP22.422.422.422.4(6.6)absent6.6(\uparrow 6.6)( ↑ 6.6 )28.328.328.328.3(0)0(0)( 0 )73.273.273.273.2(0.3)absent0.3(\uparrow 0.3)( ↑ 0.3 )43.143.143.143.1(8.0)absent8.0(\uparrow 8.0)( ↑ 8.0 )25.425.425.425.4(2.8)absent2.8(\downarrow 2.8)( ↓ 2.8 )38.538.538.538.5(2.4)absent2.4(\uparrow 2.4)( ↑ 2.4 )
24.024.024.024.0(8.2)absent8.2(\uparrow 8.2)( ↑ 8.2 )28.728.728.728.7(0.4)absent0.4(\uparrow 0.4)( ↑ 0.4 )72.272.272.272.2(0.7)absent0.7(\downarrow 0.7)( ↓ 0.7 )51.751.751.751.7(16.6)absent16.6(\uparrow 16.6)( ↑ 16.6 )26.826.826.826.8(1.4)absent1.4(\downarrow 1.4)( ↓ 1.4 )40.740.740.740.7(4.6)absent4.6(\uparrow 4.6)( ↑ 4.6 )
+In-context SP24.324.324.324.3(8.5)absent8.5(\uparrow 8.5)( ↑ 8.5 )30.330.330.330.3(2.0)absent2.0(\uparrow 2.0)( ↑ 2.0 )73.973.973.973.9(1.0)absent1.0(\uparrow 1.0)( ↑ 1.0 )50.650.650.650.6(15.5)absent15.5(\uparrow 15.5)( ↑ 15.5 )29.429.429.429.4(0.8)absent0.8(\uparrow 0.8)( ↑ 0.8 )41.741.741.741.7(5.6)absent5.6(\uparrow 5.6)( ↑ 5.6 )
25.325.325.325.3(9.5)absent9.5(\uparrow 9.5)( ↑ 9.5 )29.529.529.529.5(1.2)absent1.2(\uparrow 1.2)( ↑ 1.2 )73.973.973.973.9(1.0)absent1.0(\uparrow 1.0)( ↑ 1.0 )52.952.952.952.9(17.8)absent17.8(\uparrow 17.8)( ↑ 17.8 )28.628.628.628.6(0.4)absent0.4(\uparrow 0.4)( ↑ 0.4 )42.042.042.042.0(5.9)absent5.9(\uparrow 5.9)( ↑ 5.9 )
+Auto-SP24.324.324.324.3(8.5)absent8.5(\uparrow 8.5)( ↑ 8.5 )29.929.929.929.9(1.6)absent1.6(\uparrow 1.6)( ↑ 1.6 )72.672.672.672.6(0.3)absent0.3(\downarrow 0.3)( ↓ 0.3 )54.054.054.054.0(18.9)absent18.9(\uparrow 18.9)( ↑ 18.9 )27.627.627.627.6(0.6)absent0.6(\downarrow 0.6)( ↓ 0.6 )41.741.741.741.7(5.6)absent5.6(\uparrow 5.6)( ↑ 5.6 )
24.324.324.324.3(8.5)absent8.5(\uparrow 8.5)( ↑ 8.5 )30.330.330.330.3(2.0)absent2.0(\uparrow 2.0)( ↑ 2.0 )72.972.972.972.9(0)0(0)( 0 )56.756.756.756.7(21.6)absent21.6(\uparrow 21.6)( ↑ 21.6 )28.228.228.228.2(0)0(0)( 0 )42.542.542.542.5(6.4)absent6.4(\uparrow 6.4)( ↑ 6.4 )
+Complex-SP23.023.023.023.0(7.2)absent7.2(\uparrow 7.2)( ↑ 7.2 )29.929.929.929.9(1.6)absent1.6(\uparrow 1.6)( ↑ 1.6 )73.273.273.273.2(0.3)absent0.3(\uparrow 0.3)( ↑ 0.3 )52.352.352.352.3(17.2)absent17.2(\uparrow 17.2)( ↑ 17.2 )29.629.629.629.6(1.4)absent1.4(\uparrow 1.4)( ↑ 1.4 )41.641.641.641.6(5.5)absent5.5(\uparrow 5.5)( ↑ 5.5 )
24.624.624.624.6(8.8)absent8.8(\uparrow 8.8)( ↑ 8.8 )28.728.728.728.7(0.4)absent0.4(\uparrow 0.4)( ↑ 0.4 )72.972.972.972.9(0)0(0)( 0 )55.755.755.755.7(20.6)absent20.6(\uparrow 20.6)( ↑ 20.6 )30.030.030.030.0(1.8)absent1.8(\uparrow 1.8)( ↑ 1.8 )42.442.442.442.4(6.3)absent6.3(\uparrow 6.3)( ↑ 6.3 )

3 Self-Polish Prompting

In this section, we first revisit previous prompting paradigms aiming at solving reasoning problems. Next, we describe the proposed Self-Polish method detailedly.

3.1 Revisiting Paradigms of Reasoning Problem Solving

In the context of enhancing the capabilities of LLMs, the prompting technique has emerged as one of the most popular approaches owing to its training-free nature and effectiveness (Qiao etal., 2022; Lewkowycz etal., 2022).Here, we formalize several representative paradigms. See Figure 1 for a schematic comparison between them and our method.

Standard.

The prompt contains k×k\timesitalic_k × [Problem, Answer] pairs, followed by the test problem.

Chain-of-Thought (Wei etal., 2022b).

The prompt contains k×k\timesitalic_k × [Problem, Rationale, Answer] tuples, followed by the test problem. This method teaches models to generate rationales and answers, achieving significant improvement in reasoning. Auto-CoT (Fu etal., 2022) and Complex-CoT Zhou etal. (2022a) are two automatic varients that constructs CoT demonstrations according to the criteria of problem diversity and reasoning complexity, respecticely.

Least-to-Most Zhou etal. (2022a).

The models are taught to first reduce problems into sub-problems and then solve them sequentially. There are two kinds of prompts. The first is the problem reduction prompt that contains m×m\timesitalic_m × [OriginalProblem, Sub-Problems] pairs, followed by the test original problem. The second is the problem-solving prompt that contains k×k\timesitalic_k × [OriginalProblem, n×n\timesitalic_n × (Sub-Problem, Sub-Answer)] tuples, followed by the test original problem and the current sub-problem to solve.

Summary.

All previous methods focus on the answer/reasoning side, and it is convenient to combine them with Self-Polish which puts emphasis on the problem side.

3.2 Problem-Refining Prompting

3.2.1 Refining Principles

We expect the newly generated problems to be easier to understand and process, so they should adhere to the following principles: (1) conciseness, the problems should not be overly long, ensuring they remain easily understandable; (2) clarity, the problems should avoid ambiguous phrasing and instead utilize quantitative representations (e.g., Arabic numerals) whenever possible; (3) focus: the problems should clearly convey the intended subject matter, making it evident what the question is asking; (4) absence of irrelevant information: the problems should be free from extraneous details that could cause confusion or distractions.

3.2.2 Construction of Refining Prompts

Zero-shot Self-Polish.

It is difficult to internalize the aforementioned principles within the model via training due to the tedious process of constructing a corresponding dataset and potential catastrophic forgetting problems (Goodfellow etal., 2014; Parisi etal., 2019). So we turn to training-free strategies.

As LLMs demonstrate emergent abilities of instruction-following (Schaeffer etal., 2023; Sanh etal., 2022; Wei etal., 2022a), a simple and intuitive strategy to refine problems is prompting LLMs with an instruction. In the instruction, we guide the model to rewrite new versions of the original reasoning problem to be more understandable and easy to answer, and never omit any useful information. The prompt contains [Instruction, OriginalProblem] and the model responds with a newly generated problem. Next, we can adopt any prompting method in Section 3.1 to get the answer to the new problem, and we take this answer as the final one. We conduct preliminary validation experiments and the results are illustrated in Table 1. Zero-shot refining can consistently improve reasoning performance on various benchmarks.

In-context Self-Polish.

As empirical results show that zero-shot refining can only provide limited performance gain, especially on difficult datasets, we then add demonstrations to the prompt to enable models to better internalize and apply design principles.Specifically, demonstrations are formulated as [OriginalProblem, New Problem] pairs, and we incorporate a curated collection of problem-refining patterns in the demonstrations: (1) remove irrelevant information, as the first iteration in Figure 2; (2) rearrange the logic structure and group relevant conditions together to better match the reasoning logic of the model, as the second iteration in Figure 2; (3) summarize local conditions into new ones in parallel, as the third iteration in Figure 2.222Note that a single example typically does not encompass all refining strategies. The example is constructed solely to illustrate our design patterns.Results in Table 1 show that in-context problem refining yields more performance gain than zero-shot refining.

Automatic Self-Polish.

This is an automatic variant of the in-context problem-refining.We draw inspiration from Zhang etal. (2022) and construct the refining prompt according to the diverse semantics of problems with the technique of k𝑘kitalic_k-means clustering.The underlying hypothesis is that a diverse set of demonstrations can cover a broad semantic space of problems, thereby the model can locate relevant reference demonstrations for more test examples.Table 1 shows that Auto-SP also yields significant improvement.

Complexity-based Self-Polish.

This is another variant of the in-context problem-refining for automatically selecting refining demonstrations. We draw inspiration from Fu etal. (2022) and construct the refining prompt according to the complexity of each problem. The underlying hypothesis is that the refining ability of the model can generalize from complex problems to simpler ones.Table 1 demonstrates that Complex-SP can also yield substantial performance gain.

3.3 Progressively Refining Framework

To enhance the consistency and reliability of the refined problems, we propose a progressive framework that has two stages: the problem-solving stage (Section 3.1) and the problem-refining stage (Section 3.2). The two stages are executed alternatively until the return condition is satisfied.

Return condition & answer selection.

There are two situations that terminate the iterative process. The first is when the last two answers are the same, indicating convergence of the answer. In this case, we can directly return the answer. The second situation is when the iteration number exceeds the maximum count T=2𝑇2T=2italic_T = 2.333One iteration means one time of problem refinement. Note that a bigger T𝑇Titalic_T can yield a larger performance gain, as discussed in Section 5.1 Here we set T=2𝑇2T=2italic_T = 2 to achieve a balance in computational efficiency and performance.In such case, we have multiple options for selecting the final answer, such asthe answer to the original problem,the answer to the first generated problem, the answer to the last generated problem, or utilizing a majority voting approach to select the answer (Wang etal., 2022), which will be discussed in our ablation study in Section 5.1. Here we choose the answer to the last generated problem by default.As shown in Table 1, adding Progressively Refining to our method can bring further improvement across different prompt-construction approaches.

The overall framework is shown in Algorithm 1 in Appendix B.

Enhance Reasoning in Large Language Models via Problem Refinement (3)
Enhance Reasoning in Large Language Models via Problem Refinement (4)
Enhance Reasoning in Large Language Models via Problem Refinement (5)
Enhance Reasoning in Large Language Models via Problem Refinement (6)
Enhance Reasoning in Large Language Models via Problem Refinement (7)
Enhance Reasoning in Large Language Models via Problem Refinement (8)

Answer SideProblem SideDatasetAverage
GSM8KAQuASVAMPMultiArithMathQA
Chain-of-ThoughtNo Refinement56.156.156.156.144.944.944.944.980.380.380.380.390.890.890.890.841.041.041.041.062.662.662.662.6
In-context SP56.756.756.756.7(0.6)absent0.6(\uparrow 0.6)( ↑ 0.6 )48.848.848.848.8(3.9)absent3.9(\uparrow 3.9)( ↑ 3.9 )81.681.681.681.6(1.3)absent1.3(\uparrow 1.3)( ↑ 1.3 )88.588.588.588.5(2.3)absent2.3(\downarrow 2.3)( ↓ 2.3 )43.043.043.043.0(2.0)absent2.0(\uparrow 2.0)( ↑ 2.0 )63.763.763.763.7(1.1)absent1.1(\uparrow 1.1)( ↑ 1.1 )
Auto-SP56.956.956.956.9(0.8)absent0.8(\uparrow 0.8)( ↑ 0.8 )49.249.249.249.2(4.3)absent4.3(\uparrow 4.3)( ↑ 4.3 )78.378.378.378.3(2.0)absent2.0(\downarrow 2.0)( ↓ 2.0 )89.789.789.789.7(1.1)absent1.1(\downarrow 1.1)( ↓ 1.1 )42.842.842.842.8(1.8)absent1.8(\uparrow 1.8)( ↑ 1.8 )63.463.463.463.4(0.8)absent0.8(\uparrow 0.8)( ↑ 0.8 )
Complex-SP58.158.158.158.1(2.0)absent2.0(\uparrow 2.0)( ↑ 2.0 )47.347.347.347.3(2.4)absent2.4(\uparrow 2.4)( ↑ 2.4 )81.381.381.381.3(1.0)absent1.0(\uparrow 1.0)( ↑ 1.0 )90.290.290.290.2(0.6)absent0.6(\downarrow 0.6)( ↓ 0.6 )43.243.243.243.2(2.2)absent2.2(\uparrow 2.2)( ↑ 2.2 )64.064.064.064.0(1.4)absent1.4(\uparrow 1.4)( ↑ 1.4 )
Least-to-MostNo Refinement59.359.359.359.342.142.142.142.182.982.982.982.983.983.983.983.939.039.039.039.061.461.461.461.4
In-context SP61.261.261.261.2(1.9)absent1.9(\uparrow 1.9)( ↑ 1.9 )44.144.144.144.1(2.0)absent2.0(\uparrow 2.0)( ↑ 2.0 )84.9(2.0)absent2.0(\uparrow 2.0)( ↑ 2.0 )85.185.185.185.1(1.2)absent1.2(\uparrow 1.2)( ↑ 1.2 )41.241.241.241.2(2.2)absent2.2(\uparrow 2.2)( ↑ 2.2 )63.363.363.363.3(1.9)absent1.9(\uparrow 1.9)( ↑ 1.9 )
Auto-SP61.661.661.661.6(2.3)absent2.3(\uparrow 2.3)( ↑ 2.3 )44.944.944.944.9(2.8)absent2.8(\uparrow 2.8)( ↑ 2.8 )82.982.982.982.9(0)0(0)( 0 )83.983.983.983.9(0)0(0)( 0 )41.241.241.241.2(2.2)absent2.2(\uparrow 2.2)( ↑ 2.2 )62.962.962.962.9(1.5)absent1.5(\uparrow 1.5)( ↑ 1.5 )
Complex-SP62.962.962.962.9(3.6)absent3.6(\uparrow 3.6)( ↑ 3.6 )47.647.647.647.6(5.5)absent5.5(\uparrow 5.5)( ↑ 5.5 )84.284.284.284.2(1.3)absent1.3(\uparrow 1.3)( ↑ 1.3 )86.286.286.286.2(2.3)absent2.3(\uparrow 2.3)( ↑ 2.3 )40.640.640.640.6(1.6)absent1.6(\uparrow 1.6)( ↑ 1.6 )64.364.364.364.3 (2.9)absent2.9(\uparrow 2.9)( ↑ 2.9 )
Auto-CoTNo Refinement59.459.459.459.446.546.546.546.575.675.675.675.692.592.592.592.541.441.441.441.463.163.163.163.1
In-context SP59.459.459.459.4(0)0(0)( 0 )47.247.247.247.2(0.7)absent0.7(\uparrow 0.7)( ↑ 0.7 )77.677.677.677.6(2.0)absent2.0(\uparrow 2.0)( ↑ 2.0 )90.690.690.690.6(1.9)absent1.9(\downarrow 1.9)( ↓ 1.9 )45.8(4.4)absent4.4(\uparrow 4.4)( ↑ 4.4 )64.164.164.164.1(1.0)absent1.0(\uparrow 1.0)( ↑ 1.0 )
Auto-SP60.560.560.560.5(1.1)absent1.1(\uparrow 1.1)( ↑ 1.1 )48.048.048.048.0(1.5)absent1.5(\uparrow 1.5)( ↑ 1.5 )75.675.675.675.6(0)0(0)( 0 )89.789.789.789.7(2.8)absent2.8(\downarrow 2.8)( ↓ 2.8 )44.444.444.444.4(3.0)absent3.0(\uparrow 3.0)( ↑ 3.0 )63.663.663.663.6(0.5)absent0.5(\uparrow 0.5)( ↑ 0.5 )
Complex-SP60.160.160.160.1(0.7)absent0.7(\uparrow 0.7)( ↑ 0.7 )50.8(4.3)absent4.3(\uparrow 4.3)( ↑ 4.3 )78.278.278.278.2(2.6)absent2.6(\uparrow 2.6)( ↑ 2.6 )93.193.193.193.1(0.6)absent0.6(\uparrow 0.6)( ↑ 0.6 )43.043.043.043.0(1.6)absent1.6(\uparrow 1.6)( ↑ 1.6 )65.065.065.065.0(1.9)absent1.9(\uparrow 1.9)( ↑ 1.9 )
Complex-CoTNo Refinement67.567.567.567.547.647.647.647.678.378.378.378.394.343.643.643.643.666.366.366.366.3
In-context SP66.266.266.266.2 (1.3)absent1.3(\downarrow 1.3)( ↓ 1.3 )50.8(3.1)absent3.1(\uparrow 3.1)( ↑ 3.1 )80.980.980.980.9(2.6)absent2.6(\uparrow 2.6)( ↑ 2.6 )90.890.890.890.8(3.5)absent3.5(\downarrow 3.5)( ↓ 3.5 )44.644.644.644.6(1.3)absent1.3(\uparrow 1.3)( ↑ 1.3 )66.766.766.766.7(0.4)absent0.4(\uparrow 0.4)( ↑ 0.4 )
Auto-SP68.7(1.2)absent1.2(\uparrow 1.2)( ↑ 1.2 )45.745.745.745.7(1.9)absent1.9(\downarrow 1.9)( ↓ 1.9 )78.378.378.378.3(0)0(0)( 0 )92.092.092.092.0(2.3)absent2.3(\downarrow 2.3)( ↓ 2.3 )42.442.442.442.4(1.2)absent1.2(\downarrow 1.2)( ↓ 1.2 )65.465.465.465.4(0.9)absent0.9(\downarrow 0.9)( ↓ 0.9 )
Complex-SP68.568.568.568.5(1.0)absent1.0(\uparrow 1.0)( ↑ 1.0 )50.450.450.450.4(2.8)absent2.8(\uparrow 2.8)( ↑ 2.8 )78.678.678.678.6(0.3)absent0.3(\uparrow 0.3)( ↑ 0.3 )93.193.193.193.1(1.2)absent1.2(\downarrow 1.2)( ↓ 1.2 )44.844.844.844.8(1.2)absent1.2(\uparrow 1.2)( ↑ 1.2 )67.167.167.167.1(0.8)absent0.8(\uparrow 0.8)( ↑ 0.8 )

4 Experiments

In this section, we conduct experiments to demonstrate the effectiveness and robustness of SP.

4.1 Experimental Setups.

Models.

We employ three GPT-series models, namely text-davinci-002, text-davinci-003, and GPT-3.5-Turbo Brown etal. (2020); Ouyang etal. (2022), as they are widely recognized and accessible to the public, ensuring reproducibility of our research.Our experiments are based on OpenAI’s API. All methods use greedy decoding (i.e., temperature=0𝑡𝑒𝑚𝑝𝑒𝑟𝑎𝑡𝑢𝑟𝑒0temperature=0italic_t italic_e italic_m italic_p italic_e italic_r italic_a italic_t italic_u italic_r italic_e = 0) for stable responses.

Datasets.

We evaluate the performance of our method on five reasoning datasets 444We also conducted preliminary experiments on the more challenge dataset MATH(Hendrycks etal., 2021), where Self-Polish also demonstrated promising results. See Appendix G for more results., including GSM8K (Cobbe etal., 2021), AQuA (Ling etal., 2017), SVAMP (Patel etal., 2021), MultiArith (Roy and Roth, 2015) and MathQA (Amini etal., 2019).The datasets are evaluated by prior studies in the field of multi-hop reasoning (Wei etal., 2022b; Fu etal., 2022; Zhou etal., 2022a).We evaluate on the whole test set of AQuA and GSM8K. For other datasets, we adopt the split from Mishra etal. (2022) or randomly select 500500500500 test instances, and perform 3 restarts for stable results.

Prompts.

For the sake of generalizability,GSM8K, SVAMP and MultiArith share the same Self-Polish prompts constructed from GSM8K; AQuA and MathQA share the same Self-Polish prompts constructed from AQuA. See Appendix F for SP prompts.The prompts for the standard few-shot prompting method are from Wei etal. (2022b). The prompts for Chain-of-thought, Least-to-Most, Auto-CoT and Complex-CoT are from previous work (Wei etal., 2022b; Zhou etal., 2022a; Zheng etal., 2023; Fu etal., 2022; Zhang etal., 2022). While prompts of LtM are not available for some datasets, we manually construct them.

See more implementation details in Appendix C.

4.2 Experimental Results

Standard few-shot setting.

Figure 3 shows the results of evaluating the performance in the standard few-shot setting.We can find that :(1) Our method consistently improves reasoning performance by a large margin across multiple models and datasets, indicating its capability to enhance model understanding of problems.(2) On relatively weaker models, automated prompting methods like Auto-CoT and Complex-CoT yield more gains compared to in-context SP. However, on stronger models, the differences in performance gain between the three approaches are not significant, revealing that the stronger models are less sensitive to prompts.

Enhance Reasoning in Large Language Models via Problem Refinement (9)
Enhance Reasoning in Large Language Models via Problem Refinement (10)
Enhance Reasoning in Large Language Models via Problem Refinement (11)
Enhance Reasoning in Large Language Models via Problem Refinement (12)
Enhance Reasoning in Large Language Models via Problem Refinement (13)
Enhance Reasoning in Large Language Models via Problem Refinement (14)
Combining Self-Polish with other prompting strategies.

Table 2 demonstrates the evaluating results when combining our method with other state-of-the-art reasoning-side promoting strategies.There are several critical and interesting observations:(1) Generally, SP yields substantial performance gains for all reasoning-side methods, revealing that when the model is able to better comprehend problems, both its step-by-step reasoning capabilities and problem decomposition abilities can be significantly enhanced.(2) Whether for the reasoning side or the problem side, the Complex-based approach performs the best. This indicates that LLMs have the ability to generalize from complex tasks to simple ones, both in terms of reasoning and problem refinement.(3) As Fu etal. (2022) stated, the average number of words in problems, i.e., GSM8K (46.9), AQuA (51.9), SVAMP (32.1), MultiArith (31.2), and MathQA (60.1), can serve as a proxy for measuring the reasoning complexity of each task.We find that the more challenging the task, the higher the improvement achieved by SP, highlighting its suitability for intricate reasoning tasks.It is noteworthy that when combined with the CoT-series methods, our approach has limited improvement on MultiArith. This could be because the task itself can already be well solved by CoT and is relatively simple.Excessive refinement of simple problems carries the risk of information loss or semantic alterations, leading to a decline in performance, as depicted in Figure 9.

Robustness evaluation.

GSM-IC (Shi etal., 2023) is an adversarial arithmetic reasoning dataset with distracting information in the problem to fool the model. So it is well-suited for evaluating the robustness of models.It has two splits: GSM-IC-2step which contains problems that require two reasoning steps to solve and GSM-IC-mstep which contains problems that require more than two reasoning steps to solve.As shown in Figure 4, our method enhances the robustness and reliability of various models across different prompting techniques, shielding them from the interference of low-quality problems.

5 Discussion

5.1 Ablation Studies

As mentioned in Section 3.3, the maximum iteration times T𝑇Titalic_T and the strategy to select the final answer if the convergence is not achieved are two main components of Self-Polish. Here we perform ablation studies on them.

Max iterating times T𝑇Titalic_T.

As shown in Figure 5(a) and Figure 5(c), for both the Standard and CoT methods, larger iteration counts lead to higher convergence accuracy (“Converge” in figures), which aligns with common knowledge and further demonstrates the effectiveness of our method: by gradually optimizing problems, we enable the model to handle them more easily.But when T𝑇Titalic_T is too big, the performance of SP may suffer a drop, indicating that excessive rewriting can lead to a decline in the quality of problems.We set T=2𝑇2T=2italic_T = 2 not only for the sake of efficiency, but also because it can achieve competitive performance especially when combined with CoT-series methods.

Final answer selection strategies.

We can easily observe that with a smaller T𝑇Titalic_T, the “Last One” strategy tends to have an advantage, while as the iteration count increases, other strategies become more effective, even outperforming “Last One”. This is intuitive as after multiple rewriting iterations, the semantic meaning of a problem may deviate significantly from the original one.

5.2 Analysis of Actual Iterating Times

Figure 5(a) and Figure 5(c) show that the actual iterating times Tactualsubscript𝑇𝑎𝑐𝑡𝑢𝑎𝑙T_{actual}italic_T start_POSTSUBSCRIPT italic_a italic_c italic_t italic_u italic_a italic_l end_POSTSUBSCRIPT does not grow significantly as the max iterating times T𝑇Titalic_T increases, revealing that SP can achieve a converged answer on most of the problems with few iterations.To verify this, we illustrate the distribution map of Tactualsubscript𝑇𝑎𝑐𝑡𝑢𝑎𝑙T_{actual}italic_T start_POSTSUBSCRIPT italic_a italic_c italic_t italic_u italic_a italic_l end_POSTSUBSCRIPT with T=5𝑇5T=5italic_T = 5 in Figure 5(b) and Figure 5(d). Tactualsubscript𝑇𝑎𝑐𝑡𝑢𝑎𝑙T_{actual}italic_T start_POSTSUBSCRIPT italic_a italic_c italic_t italic_u italic_a italic_l end_POSTSUBSCRIPT exhibits a long-tail distribution, with only a few samples exceeding the max times.This finding provides evidence that our method is highly efficient that consumes few additional computational resources.

5.3 Further Improvement for Self-Consistency

Self-Consistency is a prompting method that samples multiple reasoning paths and generates a consistent answer by majority vote strategy (Wang etal., 2022).It has proved effective in various reasoning benchmarks (Wang etal., 2022).Here, we combine the Self-Polish and Self-Consistency methods to investigate whether there will be further performance improvement. We conduct experiments on two difficult datasets (i.e., GSM8K and AQuA) with temperature=0.7𝑡𝑒𝑚𝑝𝑒𝑟𝑎𝑡𝑢𝑟𝑒0.7temperature=0.7italic_t italic_e italic_m italic_p italic_e italic_r italic_a italic_t italic_u italic_r italic_e = 0.7 for diversity following (Wang etal., 2022).

Results in Table 3 demonstrate that SP provides a substantial performance gain for SC in Auto-CoT and Complex-CoT manners.Moreover, an increase in the number of reasoning paths leads to a corresponding improvement in performance, showing the advantage of voting strategy.

MethodSC PathSelf-PolishDataset
AQuAMathQA
Auto-CoT561.861.861.861.865.265.265.265.2
566.166.166.166.168.468.468.468.4
1061.461.461.461.467.467.467.467.4
1065.065.065.065.071.071.071.071.0
2063.063.063.063.069.069.069.069.0
2068.172.0
Complex-CoT561.461.461.461.461.661.661.661.6
565.065.065.065.064.464.464.464.4
1065.465.465.465.462.462.462.462.4
1066.166.166.166.164.864.864.864.8
2065.465.465.465.464.064.064.064.0
2067.065.6

Enhance Reasoning in Large Language Models via Problem Refinement (15)

5.4 Case Study

To further demonstrate the effectiveness of the problem-refining patterns we proposed and how our method embodies the proposed principles, we conducted a case study as shown in Figure 6. More cases can be found in the Appendix D (Figure 7 and Figure 8).

From Figure 6, we observe that removing irrelevant information (i.e., “Grover’s neighbor made a salary of $10 last year.”) can help the model avoid distractions and facilitate accurate reasoning.Next, rearranging the problem conditions and grouping pertinent conditions together can facilitate the model in generating more effective novel deductions during the process of reasoning (e.g., resulting in the streamlined computation of the total number of face masks in Refined Problem22{}_{2}start_FLOATSUBSCRIPT 2 end_FLOATSUBSCRIPT).

Additionally, summarizing local conditions into new ones can effectively simplify complex problems, enabling the model to handle them with greater ease. This is demonstrated in the first iteration of Figure 7 and the second iteration of Figure 8.Furthermore, the second iteration in Figure 7 highlights how our approach can explicitly and precisely define the problem in a formal manner. Specifically, in the Refined Problem22{}_{2}start_FLOATSUBSCRIPT 2 end_FLOATSUBSCRIPT of Figure 7, the model accurately identifies the two teams as “Team A” and “Team B” instead of referring to them as “one team” and “the other team”, and then it is able to clearly specify the exact question to be asked. This significantly reduces the model’s burden of understanding during the reasoning process, enhancing its overall performance.

6 Conclusion

This paper focuses on a previously neglected aspect, namely the optimization of problem formulation, within the context of enhancing multi-step reasoning in large language models. We present a novel prompting method called Self-Polish which progressively refines the given reasoning problems to facilitate model comprehension and processing. It demonstrates impressive effectiveness, robustness, and reliability in various benchmarks across different models, and can seamlessly integrate with other state-of-the-art methods. We hope it could motivate future research in this field.

Limitations

Despite the significant enhancement in the reasoning performance achieved by our approach, this work still has limitations. Firstly, our criterion for convergence is based on obtaining two identical answers rather than assessing whether the problem itself has been sufficiently optimized. Future work could involve designing methods that enable the model to autonomously determine whether a problem has reached its optimal form.Secondly, we have explored two approaches to automatically construct problem-refining prompts (i.e., Auto-Sp and Complex-SP). However, in the future, it would be beneficial to incorporate more techniques for automatically generating instructions or selecting demonstrations.Thirdly, although our designed patterns for problem refining have proven highly effective, they do not encompass all possible scenarios in the real world. In the future, it is conceivable to incorporate additional patterns to further expand the scope of applicability.

Acknowledgements

The authors wish to thank the anonymous reviewers for their helpful comments. This work was partially funded by National Natural Science Foundation of China (No.62206057,61976056,62076069), Shanghai Rising-Star Program (23QA1400200), Natural Science Foundation of Shanghai (23ZR1403500), Program of Shanghai Academic Research Leader under grant 22XD1401100.

References

  • Amini etal. (2019)Aida Amini, Saadia Gabriel, Shanchuan Lin, Rik Koncel-Kedziorski, Yejin Choi,and Hannaneh Hajishirzi. 2019.Mathqa: Towardsinterpretable math word problem solving with operation-based formalisms.In Proceedings of the 2019 Conference of the North AmericanChapter of the Association for Computational Linguistics: Human LanguageTechnologies, NAACL-HLT 2019, Minneapolis, MN, USA, June 2-7, 2019, Volume1 (Long and Short Papers), pages 2357–2367. Association for ComputationalLinguistics.
  • Bommasani etal. (2021)Rishi Bommasani, DrewA. Hudson, Ehsan Adeli, RussB. Altman, Simran Arora,Sydney von Arx, MichaelS. Bernstein, Jeannette Bohg, Antoine Bosselut, EmmaBrunskill, Erik Brynjolfsson, Shyamal Buch, Dallas Card, Rodrigo Castellon,NiladriS. Chatterji, AnnieS. Chen, Kathleen Creel, JaredQuincy Davis,Dorottya Demszky, Chris Donahue, Moussa Doumbouya, Esin Durmus, StefanoErmon, John Etchemendy, Kawin Ethayarajh, LiFei-Fei, Chelsea Finn, TrevorGale, Lauren Gillespie, Karan Goel, NoahD. Goodman, Shelby Grossman, NeelGuha, Tatsunori Hashimoto, Peter Henderson, John Hewitt, DanielE. Ho, JennyHong, Kyle Hsu, Jing Huang, Thomas Icard, Saahil Jain, Dan Jurafsky,Pratyusha Kalluri, Siddharth Karamcheti, Geoff Keeling, Fereshte Khani, OmarKhattab, PangWei Koh, MarkS. Krass, Ranjay Krishna, Rohith Kuditipudi, andetal. 2021.On the opportunities andrisks of foundation models.CoRR, abs/2108.07258.
  • Brown etal. (2020)TomB. Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared Kaplan,Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, AmandaAskell, Sandhini Agarwal, Ariel Herbert-Voss, Gretchen Krueger, TomHenighan, Rewon Child, Aditya Ramesh, DanielM. Ziegler, Jeffrey Wu, ClemensWinter, Christopher Hesse, Mark Chen, Eric Sigler, Mateusz Litwin, ScottGray, Benjamin Chess, Jack Clark, Christopher Berner, Sam McCandlish, AlecRadford, Ilya Sutskever, and Dario Amodei. 2020.Language models are few-shot learners.In Advances in Neural Information Processing Systems 33: AnnualConference on Neural Information Processing Systems 2020, NeurIPS 2020,December 6-12, 2020, virtual.
  • Chowdhery etal. (2022)Aakanksha Chowdhery, Sharan Narang, Jacob Devlin, Maarten Bosma, Gaurav Mishra,Adam Roberts, Paul Barham, HyungWon Chung, Charles Sutton, SebastianGehrmann, Parker Schuh, Kensen Shi, Sasha Tsvyashchenko, Joshua Maynez,Abhishek Rao, Parker Barnes, YiTay, Noam Shazeer, Vinodkumar Prabhakaran,Emily Reif, Nan Du, Ben Hutchinson, Reiner Pope, James Bradbury, JacobAustin, Michael Isard, Guy Gur-Ari, Pengcheng Yin, Toju Duke, AnselmLevskaya, Sanjay Ghemawat, Sunipa Dev, Henryk Michalewski, Xavier Garcia,Vedant Misra, Kevin Robinson, Liam Fedus, Denny Zhou, Daphne Ippolito, DavidLuan, Hyeontaek Lim, Barret Zoph, Alexander Spiridonov, Ryan Sepassi, DavidDohan, Shivani Agrawal, Mark Omernick, AndrewM. Dai,ThanumalayanSankaranarayana Pillai, Marie Pellat, Aitor Lewkowycz, EricaMoreira, Rewon Child, Oleksandr Polozov, Katherine Lee, Zongwei Zhou, XuezhiWang, Brennan Saeta, Mark Diaz, Orhan Firat, Michele Catasta, Jason Wei,Kathy Meier-Hellstern, Douglas Eck, Jeff Dean, Slav Petrov, and NoahFiedel. 2022.Palm: Scalinglanguage modeling with pathways.CoRR, abs/2204.02311.
  • Chu etal. (2020)Zewei Chu, Mingda Chen, Jing Chen, Miaosen Wang, Kevin Gimpel, Manaal Faruqui,and Xiance Si. 2020.How toask better questions? A large-scale multi-domain dataset for rewritingill-formed questions.In The Thirty-Fourth AAAI Conference on ArtificialIntelligence, AAAI 2020, The Thirty-Second Innovative Applications ofArtificial Intelligence Conference, IAAI 2020, The Tenth AAAI Symposiumon Educational Advances in Artificial Intelligence, EAAI 2020, New York,NY, USA, February 7-12, 2020, pages 7586–7593. AAAI Press.
  • Chung etal. (2022)HyungWon Chung, LeHou, Shayne Longpre, Barret Zoph, YiTay, William Fedus,Eric Li, Xuezhi Wang, Mostafa Dehghani, Siddhartha Brahma, Albert Webson,ShixiangShane Gu, Zhuyun Dai, Mirac Suzgun, Xinyun Chen, AakankshaChowdhery, Sharan Narang, Gaurav Mishra, Adams Yu, VincentY. Zhao, YanpingHuang, AndrewM. Dai, Hongkun Yu, Slav Petrov, EdH. Chi, Jeff Dean, JacobDevlin, Adam Roberts, Denny Zhou, QuocV. Le, and Jason Wei. 2022.Scalinginstruction-finetuned language models.CoRR, abs/2210.11416.
  • Cobbe etal. (2021)Karl Cobbe, Vineet Kosaraju, Mohammad Bavarian, Jacob Hilton, Reiichiro Nakano,Christopher Hesse, and John Schulman. 2021.Training verifiers to solvemath word problems.CoRR, abs/2110.14168.
  • Creswell etal. (2022)Antonia Creswell, Murray Shanahan, and Irina Higgins. 2022.Selection-inference: Exploiting large language models for interpretablelogical reasoning.CoRR, abs/2205.09712.
  • Faruqui and Das (2018)Manaal Faruqui and Dipanjan Das. 2018.Identifying well-formednatural language questions.In Proceedings of the 2018 Conference on Empirical Methods inNatural Language Processing, Brussels, Belgium, October 31 - November 4,2018, pages 798–803. Association for Computational Linguistics.
  • Fu etal. (2022)Yao Fu, Hao Peng, Ashish Sabharwal, Peter Clark, and Tushar Khot. 2022.Complexity-basedprompting for multi-step reasoning.CoRR, abs/2210.00720.
  • Goodfellow etal. (2014)IanJ. Goodfellow, Mehdi Mirza, Xia Da, AaronC. Courville, and Yoshua Bengio.2014.An empirical investigation ofcatastrophic forgeting in gradient-based neural networks.In 2nd International Conference on Learning Representations,ICLR 2014, Banff, AB, Canada, April 14-16, 2014, Conference TrackProceedings.
  • Hendrycks etal. (2021)Dan Hendrycks, Collin Burns, Saurav Kadavath, Akul Arora, Steven Basart, EricTang, Dawn Song, and Jacob Steinhardt. 2021.Measuring mathematical problem solving with the MATH dataset.In Proceedings of the Neural Information Processing SystemsTrack on Datasets and Benchmarks 1, NeurIPS Datasets and Benchmarks 2021,December 2021, virtual.
  • Kojima etal. (2022)Takeshi Kojima, ShixiangShane Gu, Machel Reid, Yutaka Matsuo, and YusukeIwasawa. 2022.Large language models are zero-shot reasoners.In NeurIPS.
  • Lewkowycz etal. (2022)Aitor Lewkowycz, Anders Andreassen, David Dohan, Ethan Dyer, HenrykMichalewski, VinayV. Ramasesh, Ambrose Slone, Cem Anil, Imanol Schlag, TheoGutman-Solo, Yuhuai Wu, Behnam Neyshabur, Guy Gur-Ari, and Vedant Misra.2022.Solving quantitative reasoning problems with language models.In NeurIPS.
  • Li etal. (2022)Yifei Li, Zeqi Lin, Shizhuo Zhang, Qiang Fu, Bei Chen, Jian-Guang Lou, andWeizhu Chen. 2022.On the advance ofmaking language models better reasoners.CoRR, abs/2206.02336.
  • Ling etal. (2017)Wang Ling, Dani Yogatama, Chris Dyer, and Phil Blunsom. 2017.Program induction byrationale generation: Learning to solve and explain algebraic word problems.In Proceedings of the 55th Annual Meeting of the Associationfor Computational Linguistics, ACL 2017, Vancouver, Canada, July 30 -August 4, Volume 1: Long Papers, pages 158–167. Association forComputational Linguistics.
  • Liu etal. (2022)Jiachang Liu, Dinghan Shen, Yizhe Zhang, Bill Dolan, Lawrence Carin, and WeizhuChen. 2022.What makes goodin-context examples for gpt-3?In Proceedings of Deep Learning Inside Out: The 3rd Workshop onKnowledge Extraction and Integration for Deep Learning Architectures,DeeLIO@ACL 2022, Dublin, Ireland and Online, May 27, 2022, pages 100–114.Association for Computational Linguistics.
  • Lu etal. (2022)Yao Lu, Max Bartolo, Alastair Moore, Sebastian Riedel, and Pontus Stenetorp.2022.Fantasticallyordered prompts and where to find them: Overcoming few-shot prompt ordersensitivity.In Proceedings of the 60th Annual Meeting of the Associationfor Computational Linguistics (Volume 1: Long Papers), ACL 2022, Dublin,Ireland, May 22-27, 2022, pages 8086–8098. Association for ComputationalLinguistics.
  • Min etal. (2022)Sewon Min, Xinxi Lyu, Ari Holtzman, Mikel Artetxe, Mike Lewis, HannanehHajishirzi, and Luke Zettlemoyer. 2022.Rethinking therole of demonstrations: What makes in-context learning work?In Proceedings of the 2022 Conference on Empirical Methods inNatural Language Processing, EMNLP 2022, Abu Dhabi, United Arab Emirates,December 7-11, 2022, pages 11048–11064. Association for ComputationalLinguistics.
  • Mishra etal. (2022)Swaroop Mishra, Matthew Finlayson, Pan Lu, Leonard Tang, Sean Welleck, ChittaBaral, Tanmay Rajpurohit, Oyvind Tafjord, Ashish Sabharwal, Peter Clark, andAshwin Kalyan. 2022.LILA: Aunified benchmark for mathematical reasoning.In Proceedings of the 2022 Conference on Empirical Methods inNatural Language Processing, EMNLP 2022, Abu Dhabi, United Arab Emirates,December 7-11, 2022, pages 5807–5832. Association for ComputationalLinguistics.
  • Otter etal. (2021)DanielW. Otter, JulianR. Medina, and JugalK. Kalita. 2021.A survey of theusages of deep learning for natural language processing.IEEE Trans. Neural Networks Learn. Syst., 32(2):604–624.
  • Ouyang etal. (2022)Long Ouyang, Jeffrey Wu, XuJiang, Diogo Almeida, CarrollL. Wainwright, PamelaMishkin, Chong Zhang, Sandhini Agarwal, Katarina Slama, Alex Ray, JohnSchulman, Jacob Hilton, Fraser Kelton, Luke Miller, Maddie Simens, AmandaAskell, Peter Welinder, PaulF. Christiano, Jan Leike, and Ryan Lowe. 2022.Training language models to follow instructions with human feedback.In NeurIPS.
  • Parisi etal. (2019)GermanIgnacio Parisi, Ronald Kemker, JoseL. Part, Christopher Kanan, andStefan Wermter. 2019.Continuallifelong learning with neural networks: A review.Neural Networks, 113:54–71.
  • Patel etal. (2021)Arkil Patel, Satwik Bhattamishra, and Navin Goyal. 2021.Are NLPmodels really able to solve simple math word problems?In Proceedings of the 2021 Conference of the North AmericanChapter of the Association for Computational Linguistics: Human LanguageTechnologies, NAACL-HLT 2021, Online, June 6-11, 2021, pages 2080–2094.Association for Computational Linguistics.
  • Qiao etal. (2022)Shuofei Qiao, Yixin Ou, Ningyu Zhang, Xiang Chen, Yunzhi Yao, Shumin Deng,Chuanqi Tan, Fei Huang, and Huajun Chen. 2022.Reasoning withlanguage model prompting: A survey.CoRR, abs/2212.09597.
  • Rae etal. (2021)JackW. Rae, Sebastian Borgeaud, Trevor Cai, Katie Millican, Jordan Hoffmann,H.Francis Song, John Aslanides, Sarah Henderson, Roman Ring, Susannah Young,Eliza Rutherford, Tom Hennigan, Jacob Menick, Albin Cassirer, Richard Powell,George vanden Driessche, LisaAnne Hendricks, Maribeth Rauh, Po-Sen Huang,Amelia Glaese, Johannes Welbl, Sumanth Dathathri, Saffron Huang, JonathanUesato, John Mellor, Irina Higgins, Antonia Creswell, Nat McAleese, Amy Wu,Erich Elsen, SiddhantM. Jayakumar, Elena Buchatskaya, David Budden, EsmeSutherland, Karen Simonyan, Michela Paganini, Laurent Sifre, Lena Martens,XiangLorraine Li, Adhiguna Kuncoro, Aida Nematzadeh, Elena Gribovskaya,Domenic Donato, Angeliki Lazaridou, Arthur Mensch, Jean-Baptiste Lespiau,Maria Tsimpoukelli, Nikolai Grigorev, Doug Fritz, Thibault Sottiaux, MantasPajarskas, Toby Pohlen, Zhitao Gong, Daniel Toyama, CypriendeMassond’Autume, Yujia Li, Tayfun Terzi, Vladimir Mikulik, IgorBabuschkin, Aidan Clark, Diego deLasCasas, Aurelia Guy, Chris Jones, JamesBradbury, MatthewJ. Johnson, BlakeA. Hechtman, Laura Weidinger, IasonGabriel, William Isaac, Edward Lockhart, Simon Osindero, Laura Rimell, ChrisDyer, Oriol Vinyals, Kareem Ayoub, Jeff Stanway, Lorrayne Bennett, DemisHassabis, Koray Kavukcuoglu, and Geoffrey Irving. 2021..CoRR, abs/2112.11446.
  • Roy and Roth (2015)Subhro Roy and Dan Roth. 2015.Solving generalarithmetic word problems.In Proceedings of the 2015 Conference on Empirical Methods inNatural Language Processing, EMNLP 2015, Lisbon, Portugal, September 17-21,2015, pages 1743–1752. The Association for Computational Linguistics.
  • Sanh etal. (2022)Victor Sanh, Albert Webson, Colin Raffel, StephenH. Bach, Lintang Sutawika,Zaid Alyafeai, Antoine Chaffin, Arnaud Stiegler, Arun Raja, Manan Dey,MSaiful Bari, Canwen Xu, Urmish Thakker, ShanyaSharma Sharma, ElizaSzczechla, Taewoon Kim, Gunjan Chhablani, NihalV. Nayak, Debajyoti Datta,Jonathan Chang, MikeTian-Jian Jiang, Han Wang, Matteo Manica, Sheng Shen,ZhengXin Yong, Harsh*t Pandey, Rachel Bawden, Thomas Wang, Trishala Neeraj,Jos Rozen, Abheesht Sharma, Andrea Santilli, Thibault Févry, JasonAlanFries, Ryan Teehan, TevenLe Scao, Stella Biderman, Leo Gao, Thomas Wolf, andAlexanderM. Rush. 2022.Multitaskprompted training enables zero-shot task generalization.In The Tenth International Conference on LearningRepresentations, ICLR 2022, Virtual Event, April 25-29, 2022.OpenReview.net.
  • Schaeffer etal. (2023)Rylan Schaeffer, Brando Miranda, and Sanmi Koyejo. 2023.Are emergentabilities of large language models a mirage?CoRR, abs/2304.15004.
  • Shi etal. (2023)Freda Shi, Xinyun Chen, Kanishka Misra, Nathan Scales, David Dohan, EdH. Chi,Nathanael Schärli, and Denny Zhou. 2023.Large languagemodels can be easily distracted by irrelevant context.CoRR, abs/2302.00093.
  • Shou and Smithson (2015)Yiyun Shou and Michael Smithson. 2015.Effects of questionformats on causal judgments and model evaluation.Frontiers in Psychology, 6.
  • Srivastava etal. (2022)Aarohi Srivastava, Abhinav Rastogi, Abhishek Rao, Abu AwalMd Shoeb, AbubakarAbid, Adam Fisch, AdamR. Brown, Adam Santoro, Aditya Gupta, AdriàGarriga-Alonso, Agnieszka Kluska, Aitor Lewkowycz, Akshat Agarwal, AletheaPower, Alex Ray, Alex Warstadt, AlexanderW. Kocurek, Ali Safaya, Ali Tazarv,Alice Xiang, Alicia Parrish, Allen Nie, Aman Hussain, Amanda Askell, AmandaDsouza, Ameet Rahane, AnantharamanS. Iyer, Anders Andreassen, AndreaSantilli, Andreas Stuhlmüller, AndrewM. Dai, Andrew La, AndrewK.Lampinen, Andy Zou, Angela Jiang, Angelica Chen, Anh Vuong, Animesh Gupta,Anna Gottardi, Antonio Norelli, Anu Venkatesh, Arash Gholamidavoodi, ArfaTabassum, Arul Menezes, Arun Kirubarajan, Asher Mullokandov, AshishSabharwal, Austin Herrick, Avia Efrat, Aykut Erdem, Ayla Karakas, and etal.2022.Beyond theimitation game: Quantifying and extrapolating the capabilities of languagemodels.CoRR, abs/2206.04615.
  • Wang etal. (2022)Xuezhi Wang, Jason Wei, Dale Schuurmans, QuocV. Le, EdH. Chi, and Denny Zhou.2022.Self-consistencyimproves chain of thought reasoning in language models.CoRR, abs/2203.11171.
  • Wei etal. (2022a)Jason Wei, Maarten Bosma, VincentY. Zhao, Kelvin Guu, AdamsWei Yu, BrianLester, Nan Du, AndrewM. Dai, and QuocV. Le. 2022a.Finetunedlanguage models are zero-shot learners.In The Tenth International Conference on LearningRepresentations, ICLR 2022, Virtual Event, April 25-29, 2022.OpenReview.net.
  • Wei etal. (2022b)Jason Wei, Xuezhi Wang, Dale Schuurmans, Maarten Bosma, Brian Ichter, Fei Xia,EdH. Chi, QuocV. Le, and Denny Zhou. 2022b.Chain-of-thought prompting elicits reasoning in large language models.In NeurIPS.
  • Xi etal. (2023)Zhiheng Xi, Wenxiang Chen, Xin Guo, Wei He, Yiwen Ding, Boyang Hong, MingZhang, Junzhe Wang, Senjie Jin, Enyu Zhou, Rui Zheng, Xiaoran Fan, Xiao Wang,Limao Xiong, Yuhao Zhou, Weiran Wang, Changhao Jiang, Yicheng Zou, XiangyangLiu, Zhangyue Yin, Shihan Dou, Rongxiang Weng, Wensen Cheng, QiZhang,Wenjuan Qin, Yongyan Zheng, Xipeng Qiu, Xuanjing Huan, and Tao Gui. 2023.The rise andpotential of large language model based agents: A survey.CoRR, abs/2309.07864.
  • Ye and Durrett (2022)XiYe and Greg Durrett. 2022.The unreliabilityof explanations in few-shot in-context learning.CoRR, abs/2205.03401.
  • Zelikman etal. (2022)Eric Zelikman, Yuhuai Wu, Jesse Mu, and NoahD. Goodman. 2022.Star: Bootstrapping reasoning with reasoning.In NeurIPS.
  • Zellers etal. (2018)Rowan Zellers, Yonatan Bisk, Roy Schwartz, and Yejin Choi. 2018.SWAG: A large-scaleadversarial dataset for grounded commonsense inference.In Proceedings of the 2018 Conference on Empirical Methods inNatural Language Processing, Brussels, Belgium, October 31 - November 4,2018, pages 93–104. Association for Computational Linguistics.
  • Zhang etal. (2022)Zhuosheng Zhang, Aston Zhang, MuLi, and Alex Smola. 2022.Automatic chain ofthought prompting in large language models.CoRR, abs/2210.03493.
  • Zheng etal. (2023)Chuanyang Zheng, Zhengying Liu, Enze Xie, Zhenguo Li, and YuLi. 2023.Progressive-hintprompting improves reasoning in large language models.CoRR, abs/2304.09797.
  • Zhou etal. (2022a)Denny Zhou, Nathanael Schärli, LeHou, Jason Wei, Nathan Scales, XuezhiWang, Dale Schuurmans, Olivier Bousquet, Quoc Le, and EdH. Chi.2022a.Least-to-mostprompting enables complex reasoning in large language models.CoRR, abs/2205.10625.
  • Zhou etal. (2022b)Hattie Zhou, Azade Nova, Hugo Larochelle, AaronC. Courville, Behnam Neyshabur,and Hanie Sedghi. 2022b.Teachingalgorithmic reasoning via in-context learning.CoRR, abs/2211.09066.

Appendix

Appendix A Disscussion of More Related Work

Recent research has unveiled an unpredictable phenomenon known as emergent abilities, which manifest exclusively in larger language models while eluding their smaller counterparts (Schaeffer etal., 2023). In-context learning, instruction following, and multi-step reasoning are three emergent abilities that we focus on. We have discussed the multi-step reasoning in Section 2 and we will discuss another two.We also compare our method with the LtM detailedly here.

In-context learning.

It is demonstrated that a large language model can learn patterns from a few input-output examples in the context (input) to perform the task for an unseen inference-time example (Brown etal., 2020; Chowdhery etal., 2022), and such ability is referred to as in-context learning (ICL).Recent studies have further highlighted the impressive performance of ICL in reasoning tasks (Wei etal., 2022b; Fu etal., 2022; Zhou etal., 2022a).In our research, we capitalize on this capability to generate new formulations of problems by injecting rephrasing patterns to the demonstrations.

Instruction following.

LLMs can learn to perform unseen tasks solely through the comprehension of task-specific natual language instructions (Sanh etal., 2022; Wei etal., 2022a; Chung etal., 2022; Ouyang etal., 2022). There is also work showing that combining instructions with in-context learning can provide further benefits and that few-shot demonstrations can be viewed as a special kind of instruction that arouses the implicit ability in LLMs (Chung etal., 2022; Zhou etal., 2022b; Qiao etal., 2022).

Compaison with LtM.

The work that is most similar to ours may be Least-to-Most (LtM) which decomposes the original problem into a series of sub-problems that need to be solved sequentially (Zhou etal., 2022a). However, LtM is an variant of CoT, and there are differences in motivation and operation process between LtM and SP. Firstly, LtM is an answer/reasoning side approach that emphasizes the decomposition of a complex problem into sub-problems, while we emphasize refining the original problem to make it more understandable.Secondly, in LtM, sub-problems are solved sequentially, requiring the answer of the previous sub-problem to tackle the next one, which can lead to fragility in the reasoning chain. In contrast, our method allows for the combination of local related conditions to form new conditions parallelly.

Appendix B The Algorithm of Self-Polish

See Algorithm 1 for the overall framework of Self-Polish.

Input: language model 𝒢𝒢\mathcal{G}caligraphic_G, problem set 𝒮𝒮{\mathcal{S}}caligraphic_S, prompt 𝒫refinesubscript𝒫𝑟𝑒𝑓𝑖𝑛𝑒\mathcal{P}_{refine}caligraphic_P start_POSTSUBSCRIPT italic_r italic_e italic_f italic_i italic_n italic_e end_POSTSUBSCRIPT of the problem side refining method, prompt 𝒫answersubscript𝒫𝑎𝑛𝑠𝑤𝑒𝑟\mathcal{P}_{answer}caligraphic_P start_POSTSUBSCRIPT italic_a italic_n italic_s italic_w italic_e italic_r end_POSTSUBSCRIPT of the answer/reasoning side method, max iteration number T𝑇Titalic_T, answer selection strategy 𝒵𝒵\mathcal{Z}caligraphic_Z.

1 for eachproblemnormal-eachnormal-problem\rm{each\ problem}roman_each roman_problem s𝑠sitalic_s innormal-in\rm{in}roman_in 𝒮𝒮{\mathcal{S}}caligraphic_Sdo

2 answer_list=[]answer_list\mathrm{answer\_list}=[\ \ ]roman_answer _ roman_list = [ ];

3 t=0𝑡0t=0italic_t = 0;

4 Procedure Generate Answer to Original Problem rationalet,anst=𝒢(𝒫answers)subscriptrationale𝑡subscriptans𝑡𝒢direct-sumsubscript𝒫𝑎𝑛𝑠𝑤𝑒𝑟𝑠\mathrm{rationale}_{t},\mathrm{ans}_{t}=\mathcal{G}(\mathcal{P}_{answer}\opluss)roman_rationale start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT , roman_ans start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT = caligraphic_G ( caligraphic_P start_POSTSUBSCRIPT italic_a italic_n italic_s italic_w italic_e italic_r end_POSTSUBSCRIPT ⊕ italic_s );

5 answer_listanswer_list\mathrm{answer\_list}roman_answer _ roman_list.append(anstsubscriptans𝑡\mathrm{ans}_{t}roman_ans start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT);

6 t=t+1𝑡𝑡1t=t+1italic_t = italic_t + 1; Procedure Iterate Problem Refinement and Answer s=𝒢(𝒫refines)𝑠𝒢direct-sumsubscript𝒫𝑟𝑒𝑓𝑖𝑛𝑒𝑠s=\mathcal{G}(\mathcal{P}_{refine}\oplus s)italic_s = caligraphic_G ( caligraphic_P start_POSTSUBSCRIPT italic_r italic_e italic_f italic_i italic_n italic_e end_POSTSUBSCRIPT ⊕ italic_s );

7 rationalet,anst=𝒢(𝒫answers)subscriptrationale𝑡subscriptans𝑡𝒢direct-sumsubscript𝒫𝑎𝑛𝑠𝑤𝑒𝑟𝑠\mathrm{rationale}_{t},\mathrm{ans}_{t}=\mathcal{G}(\mathcal{P}_{answer}\opluss)roman_rationale start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT , roman_ans start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT = caligraphic_G ( caligraphic_P start_POSTSUBSCRIPT italic_a italic_n italic_s italic_w italic_e italic_r end_POSTSUBSCRIPT ⊕ italic_s );

8 ifanst==anst1\mathrm{ans}_{t}==\mathrm{ans}_{t-1}roman_ans start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT = = roman_ans start_POSTSUBSCRIPT italic_t - 1 end_POSTSUBSCRIPTthen

9 Return anstsubscriptans𝑡\mathrm{ans}_{t}roman_ans start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT.

10 else ift>T𝑡𝑇t>Titalic_t > italic_Tthen

11 Return 𝒵𝒵\mathcal{Z}caligraphic_Z(answer_listanswer_list\mathrm{answer\_list}roman_answer _ roman_list).

12 else

13 answer_listanswer_list\mathrm{answer\_list}roman_answer _ roman_list.append(anstsubscriptans𝑡\mathrm{ans}_{t}roman_ans start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT);

14 t=t+1𝑡𝑡1t=t+1italic_t = italic_t + 1;

15 end if

16

17 end for

Appendix C Implementation Details

We set the maximum iterating count to T=2𝑇2T=2italic_T = 2. Note that the bigger maximum iteration count T𝑇Titalic_T may lead to better performance, but here we set it to 2222 to achieve a trade-off between computational efficiency and effectiveness.

When combining with other reasoning-side methods (i.e., CoT, LtM, Complex-CoT and Auto-CoT) on MultiArith and SVAMP, we set the answer selection strategy as “selecting the answer to the original problem” because this dataset is relatively easy for these prompting methods. Actually, in cases where it is not necessary, rewriting easy problems may result in the loss of critical information or altering the semantics of the original problem.In other settings, we set the answer selection strategy as “selecting the answer to the last problem”.

Appendix D More Cases and Examples

Here we list more cases of Self-Polish in Figure 7 and Figure 8. We also list the failure case of excessive problem refining in Figure 9

Enhance Reasoning in Large Language Models via Problem Refinement (16)
Enhance Reasoning in Large Language Models via Problem Refinement (17)
Enhance Reasoning in Large Language Models via Problem Refinement (18)

Appendix E Sensitivity to Number and Order of Demonstrations

MethodShotsMeanOrder Deviation
Std+SP222221.321.321.321.30.80.80.80.8
333322.522.522.522.50.60.60.60.6
444424.124.124.124.10.90.90.90.9
555525.025.025.025.01.51.51.51.5
666626.326.326.326.30.90.90.90.9
CoT+SP222259.059.059.059.02.02.02.02.0
333361.361.361.361.31.51.51.51.5
444461.561.561.561.51.11.11.11.1
555562.162.162.162.11.81.81.81.8
666661.361.361.361.31.61.61.61.6

As widely recognized, in-context learning is highly sensitive to the number and order of demonstrations within the prompt (Min etal., 2022; Lu etal., 2022; Liu etal., 2022). In this regard, we investigate whether our problem-refining process is sensitive to these variables via experiments on GSM8K with Text-davinci-003. We randomly select 200200200200 examples from the test set. For a specific shot number, we randomly select five sets of demonstrations. For each set of demonstrations, we obtain performance results in five different orders.We observed that in the standard manner, increasing the number of demonstrations leads to improved performance. However, in the CoT manner, the performance converges when the number of shots is equal to 5555, demonstrating impressive sample efficiency. Additionally, in the standard manner, our method is not sensitive to the order of demonstrations while it is highly sensitive to the order of demonstrations in the CoT manner.

Appendix F Prompts of Self-Polish

The in-context Self-Polish prompt for AQuA and MathQA is in Table LABEL:table:In-context_SP_prompt_for_AQuA_and_MathQA.. The Auto-SP prompt for AQuA and MathQA is in Table LABEL:table:Auto-SP_prompt_for_AQuA_and_MathQA. and Table LABEL:table:Continuation_of_Auto-SP_prompt_for_AQuA_and_MathQA.. The Complex-SP prompt for AQuA and MathQA is in Table LABEL:table:Complex-SP_prompt_for_AQuA_and_MathQA. and Table LABEL:table:Continuation_of_Complex-SP_prompt_for_AQuA_and_MathQA..

The in-context Self-Polish prompt for GSM8K, SVAMP and MultiArith is in Table LABEL:table:In-context_SP_prompt_for_GSM8K,_SVAMP_and_MultiArith.. The Auto-SP prompt for GSM8K, SVAMP and MultiArith is in Table LABEL:table:Auto-SP_prompt_for_GSM8K,_SVAMP_and_MultiArith. and Table LABEL:table:Continuation_of_Auto-SP_prompt_for_GSM8K,_SVAMP_and_MultiArith.. The Complex-SP prompt for GSM8K, SVAMP and MultiArith is in Table LABEL:table:Complex-SP_prompt_for_GSM8K,_SVAMP_and_MultiArith. and Table LABEL:table:Continuation_of_Complex-SP_prompt_for_GSM8K,_SVAMP_and_MultiArith..

Appendix G More results on MATH dataset

As Table 5 shows, we also conducted Self-Polish methods on the MATH dataset (Hendrycks etal., 2021). Our approach demonstrated promising results. Specifically, we randomly selected 200 samples for testing, and use the Chain-of-Thought as the answer-side method.

MethodMATH
No Refinement21212121
Zero-shot SP23.523.523.523.5
In-context SP24.524.524.524.5

Please rewrite new versions of the original mathematical question (including the context and the final question) to be more understandable and easy to answer. Don’t omit any useful information, especially the numbers.

Original Question: Krishan and Nandan jointly started a business. Krishan invested six times as Nandan did and invested his money for double time as compared to Nandan. Nandan earned Rs. 6000. If the gain is proportional to the money invested and the time for which the money is invested then the total gain was? Answer Choices: (A) Rs.78000 (B) Rs.48000 (C) Rs.6000 (D) Rs.82000 (E) Rs.32000

New Question: Krishan and Nandan teamed up to start a business together. Krishan invested 12 times more money than Nandan did. Nandan’s earnings from the business were Rs. 6000. If the gain is directly proportional to both the amount of money invested and the time period, what was the total gain for both of them? Answer Choices: (A) Rs.78000 (B) Rs.48000 (C) Rs.6000 (D) Rs.82000 (E) Rs.32000

Original Question: In a graduate physics course, 70 percent of the students are male and 30 percent of the students are married. If two-sevenths of the male students are married, what fraction of the male students is single? Answer Choices: (A) 2/7 (B) 1/3 (C) 1/2 (D) 2/3 (E) 5/7

New Question: In a graduate physics course, 7/10 of the students are male and 3/10 of the students are married. If 2/7 of the male students are married, what fraction of the male students is single? Answer Choices: (A) 2/7 (B) 1/3 (C) 1/2 (D) 2/3 (E) 5/7

Original Question: A train 500m long can cross an electric pole in 20 sec and then find the speed of the train? Answer Choices: (A) 95 Kmph (B) 90 Kmph (C) 92 Kmph (D) 95 Kmph (E) 98 Kmph

New Question: A train, which is 500 meters long, takes 20 seconds to pass by an electric pole. What is the speed of the train? Represent the answer in units from answer options. Answer Choices: (A) 95 Kmph (B) 90 Kmph (C) 92 Kmph (D) 95 Kmph (E) 98 Kmph

Original Question: A train covers a distance of 10km in 10 min. If it takes 6 sec to pass a telegraph post, then the length of the train is? Answer Choices: (A) 50m (B) 60m (C) 100m (D) 90m (E) 120m

New Question: A train covers a distance of 10000m in 600 sec. If it takes 6 sec to pass a telegraph post, then the length of the train is? Answer Choices: (A) 50m (B) 60m (C) 100m (D) 90m (E) 120m

Original Question: How many different subsets of the set {0, 1, 2, 3, 4} do not contain 0? Answer Choices: (A) 16 (B) 27 (C) 31 (D) 32 (E) 64

New Question: How many different subsets of the set { 1, 2, 3, 4} ? Answer Choices: (A) 16 (B) 27 (C) 31 (D) 32 (E) 64

Original Question: A class has 6 boys and x girls. Average score of boys and girls is 50 and 60 respectively. the average of the whole class is 55, what is the value of x? Answer Choices: (A) 5 (B) 6 (C) 10 (D) 12 (E) 15

New Question: In a class, there are 6 boys and an unknown number of girls. The average score of the boys is 50, while the average score of the girls is 60. The overall average score of the boys and girls is 55. How many girls are there in the class? Answer Choices: (A) 5 (B) 6 (C) 10 (D) 12 (E) 15

Please rewrite new versions of the original mathematical question (including the context and the final question) to be more understandable and easy to answer. Don’t omit any useful information, especially the numbers.

Original Question: A and B can together finish a work in 40days. They worked together for 10days and then B left. After another 12days, A finished the remaining work. In how many days A alone can finish the job? Answer Choices: (A) 10 (B) 25 (C) 60 (D) 16 (E) 20

New Question: A and B can together finish a work in 40 days. They worked together for 10 days and then B left, and the remaining work is 3/4 of the original one. After another 12 days, A finished the remaining work alone. In how many days A alone can finish the whole job? Answer Choices: (A) 10 (B) 25 (C) 60 (D) 16 (E) 20

Original Question: A man buys an article and sells it at a profit of 20%. If he had bought it at 20% less and sold it for Rs.75 less, he could have gained 25%. What is the cost price? Answer Choices: (A) 388 (B) 375 (C) 288 (D) 266 (E) 269

New Question: A man buys an article at the price of x and sold it at the price of 1.2x, if he had bought it at a 20% discount which is 0.8x and sold it for Rs.75 less than 1.2x, he would have gained 25% of 0.8x. What was the original price of the article before any discounts or markups? Answer Choices: (A) 388 (B) 375 (C) 288 (D) 266 (E) 269

Original Question: The numbers of students speaking English and Hindi are in the ratio of 4 : 5. If the number of students speaking English increased by 35% and that speaking Hindi increased by 20%, what would be the new respective ratio? Answer Choices: (A) 19 : 20 (B) 7 : 8 (C) 8 : 9 (D) Cannot be determined (E) None of these

New Question: The number of students speaking English is 400 and increased by 35%. The number of students speaking Hindi is 500 and increased by 20%, what would be the respective ratio? what is the new ratio of students speaking English and Hindi? Answer Choices: (A) 19 : 20 (B) 7 : 8 (C) 8 : 9 (D) Cannot be determined (E) None of these

Original Question: A rectangular field has area equal to 150 sq m and perimeter 50 m. Its length and breadth must be? Answer Choices: (A) 10 (B) 88 (C) 66 (D) 65 (E) 22

New Question: Let l and b be the length and the breadth of the rectangular. The area of a rectangular field is 150 square meters: l*b = 50, and its perimeter is 50 meters: 2l + 2b =50. What are the breadth of the field? Answer Choices: (A) 10 (B) 88 (C) 66 (D) 65 (E) 22

Original Question: The ratio of two numbers is 3:4 and their sum is 14. The greater of the two numbers is? Answer Choices: (A) 12 (B) 14 (C) 16 (D) 8 (E) 19

New Question: There are two number a and b. The sum of a and b is 14, and the ratio of a and b 3:4. What is b? Answer Choices: (A) 12 (B) 14 (C) 16 (D) 8 (E) 19

Original Question: A and B invests Rs.6000 and Rs.8000 in a business. After 6 months, A withdraws half of his capital and B withdraws one-fourth of his capital. In what ratio should they share the profits at the end of the year? Answer Choices: (A) 13:15 (B) 9:13 (C) 9:11 (D) 13:14 (E) 9:14

New Question: A and B invested Rs.6000 and Rs.8000 respectively in a business. After 6 months,A withdraws half of his investment and B withdraws 1/4 of his investment. What is the ratio of their remaining investment? Answer Choices: (A) 13:15 (B) 9:13 (C) 9:11 (D) 13:14 (E) 9:14

Original Question: A train 640 meters long is running with a speed of 64 kmph. The time taken by it to cross a tunnel 140 meters long is? Answer Choices: (A) 44 sec (B) 49 sec (C) 48 sec (D) 16 sec (E) 17 sec

New Question: A train is running with a speed of 64kmph. The length of train is 640 meters and there is a tunnel 140 meters long. The time taken by the train to cross tunnel is? Answer Choices: (A) 44 sec (B) 49 sec (C) 48 sec (D) 16 sec (E) 17 sec

Original Question: There are 15 boys and 10 girls in a class. If three students are selected at random, in how many ways that 1 girl and 2 boys are selected ? Answer Choices: (A) 950 (B) 1050 (C) 2150 (D) 2050 (E) 1000

New Question: There are 15 boys and 10 girls in a class. If three students are selected at random, how many total ways can 1 girl be chosen in 10 girls and 2 boys be chosen in 15 boys? Answer Choices: (A) 950 (B) 1050 (C) 2150 (D) 2050 (E) 1000

Please rewrite new versions of the original mathematical question (including the context and the final question) to be more understandable and easy to answer. Don’t omit any useful information, especially the numbers.

Original Question: There were 35 students in a hostel. Due to the admission of 7 new students the expenses of the mess were increased by rs .84 per day while the average expenditure per head diminished by re 1. What was the original expenditure of the mess? Answer Choices: (A) rs 450 (B) rs 920 (C) rs 550 (D) rs . 630 (E) none of these

New Question: In a hostel, there were initially 35 students, and then 7 new students were admitted. While the average expenditure per student decreased by Re. 1, the daily expenses of the new mess increased by Rs. 84. What was the original daily expenditure of the mess? Answer Choices: (A) rs 450 (B) rs 920 (C) rs 550 (D) rs . 630 (E) none of these

Original Question: A train 200m long passes a man , running at 5km / hr in the same direction in which the train is going, in 10 seconds. The speed of the train is? Answer Choices: (A) 28 (B) 50 (C) 77 (D) 22 (E) 12

New Question: A train, which is 200 meters long, passes a man running at 5 kilometers per hour in the same direction as the train in 10 seconds. What is the speed of the train? Answer Choices: (A) 28 (B) 50 (C) 77 (D) 22 (E) 12

Original Question: Solution x contains 20 % of material a and 80 % of material b . solution y contains 30 % of material a and 70 % of material b . a mixture of both these solutions contains 22 % of material a in the final product . How much solution x is present in the mixture ? Answer Choices: (A) 40 % (B) 60 % (C) 80 % (D) 100 % (E) 110 %

New Question: A mixture of solution x and y contains 22 % of material a. Solution x contains 20 % of material a and 80 % of material b while solution y contains 30 % of material a and 70 % of material b. What percentage of solution x is present in the mixture? Answer Choices: (A) 40 % (B) 60 % (C) 80 % (D) 100 % (E) 110 %

Original Question: A trader sells 40 metres of cloth for rs.8200 at a profit of rs.35 per metre of cloth. How much profit will the trader earn on 40 metres of cloth ? Answer Choices: (A) rs.950 (B) rs . 1500 (C) rs . 1000 (D) rs . 1400 (E) none of these

New Question: A trader sells 40 meters of cloth and makes a profit of Rs. 35 per meter of cloth. How much profit does the trader make from selling 40 meters of cloth? Answer Choices: (A) rs . 950 (B) rs . 1500 (C) rs . 1000 (D) rs . 1400 (E) none of these

Original Question: If x < y < z and y - x > 5 , where x is an even integer and y and z are odd integers , what is the least possible value s of z - x ? Answer Choices: (A) 6 (B) 7 (C) 8 (D) 9 (E) 10

New Question: If x is an even integer, y and z are odd integers, and y is greater than x by more than 5, and z is greater than y. What is the smallest possible difference between z and x? Answer Choices: (A) 6 (B) 7 (C) 8 (D) 9 (E) 10

Original Question: What is the difference between the c.i. on rs . 6000 for 1 1/2 years at 4 % per annum compounded yearly and half-yearly? Answer Choices: (A) s.2.04 (B) s.2.08 (C) s.2.02 (D) s.2.83 (E) s.2.45

New Question: What is the difference in the compound interest earned on Rs. 6000 for 1.5 years at 4% per annum when compounded yearly and when compounded half-yearly? Answer Choices: (A) s.2.04 (B) s.2.08 (C) s.2.02 (D) s.2.83 (E) s.2.45

Original Question: The average weight of a, b and c is 45 kg. If the average weight of a and b be 40 kg and that of b and c be 45 kg , then the weight of b is? Answer Choices: (A) 31 kg (B) 32 kg (C) 33 kg (D) 35 kg (E) none of these

New Question: The average weight of a, b and c is 45 kg, which means the total weight of a, b and c is 135 kg. If the average weight of a and b is 40 kg, which means the total weight of a and b is 80kg, so the weight of c is 45kg. The average weight of b and c is 45 kg which means the total weight of b and c is 90kg. What is the weight of b? Answer Choices: (A) 31 kg (B) 32 kg (C) 33 kg (D) 35 kg (E) none of these

Original Question: The compound and the simple interests on a certain sum at the same rate of interest for two years are rs.11730 and rs.10200 respectively . The sum is? Answer Choices: (A) rs.17037 (B) rs.17000 (C) rs.17276 (D) rs.170287 (E) rs.171881

New Question: A sum of money earns compound interest and simple interest at the same rate for two years. The compound interest is Rs.11730 and the simple interest is Rs.10200. What is the sum of money? Answer Choices: (A) rs.17037 (B) rs.17000 (C) rs.17276 (D) rs.170287 (E) rs.171881

Please rewrite new versions of the original mathematical question (including the context and the final question) to be more understandable and easy to answer. Don’t omit any useful information, especially the numbers.

Original Question: Each bird eats 12 beetles per day, each snake eats 3 birds per day, and each jaguar eats 5 snakes per day. If there are 6 jaguars in a forest, how many beetles are eaten each day?

New Question: In a forest, there are 6 jaguars that each eat 5 snakes per day. Each snake eats 3 birds per day, and each bird eats 12 beetles per day. How many beetles are eaten each day by the jaguars?

Original Question: Albert is wondering how much pizza he can eat in one day. He buys 2 large pizzas and 2 small pizzas. A large pizza has 16 slices and a small pizza has 8 slices. If he eats it all, how many pieces does he eat that day?

New Question: Albert has purchased 2 large pizzas and 2 small pizzas and is wondering how many slices he can eat in one day. Each large pizza has 16 slices and each small pizza has 8 slices. If Albert eats all of the pizza, how many slices will he have eaten in one day?

Original Question: In a truck, there are 26 pink hard hats, 15 green hard hats, and 24 yellow hard hats. If Carl takes away 4 pink hard hats, and John takes away 6 pink hard hats and twice as many green hard hats as the number of pink hard hats that he removed, then calculate the total number of hard hats that remained in the truck.

New Question: In a truck, there are 26 pink hard hats, 15 green hard hats, and 24 yellow hard hats. Carl takes away 4 pink hard hats and John takes away 6 pink hard hats and 12 green hard hats. How many hard hats remain in the truck?

Original Question: Jasper will serve charcuterie at his dinner party. He buys 2 pounds of cheddar cheese for $10, a pound of cream cheese that cost half the price of the cheddar cheese, and a pack of cold cuts that cost twice the price of the cheddar cheese. How much does he spend on the ingredients?

New Question: Jasper is hosting a dinner party and wants to serve charcuterie. He buys 2 pounds of cheddar cheese for $10, 1 pound of cream cheese for $5, and a pack of cold cuts for $20. How much does he spend on the ingredients for the charcuterie?

Original Question: Tomas ate 1.5 pounds of chocolate fudge last week. Katya ate half a pound of peanut butter fudge, while Boris ate 2 pounds of fudge. How many ounces of fudge did the Tomas, Katya and Boris eat in total?

New Question: Tomas ate 24 ounces of chocolate fudge last week. Katya ate 8 ounces of peanut butter fudge, while Boris ate 32 ounces of fudge. How many ounces of fudge did the Tomas, Katya and Boris eat in total?

Original Question: Tomas ate 24 ounces of chocolate fudge last week. Katya ate 8 ounces of peanut butter fudge, while Boris ate 32 ounces of fudge. How many ounces of fudge did the Tomas, Katya and Boris eat in total?

New Question: Tomas ate 24 ounces of fudge last week. Katya ate 8 ounces of fudge, while Boris ate 32 ounces of fudge. How many ounces of fudge did the Tomas, Katya and Boris eat in total?

Please rewrite new versions of the original mathematical question (including the context and the final question) to be more understandable and easy to answer. Don’t omit any useful information, especially the numbers.

Original Question: Monica is a teacher. She has 6 classes per day. The first class has 20 students. The second and third classes have 25 students. Her fourth class has half as many as her first class. Her fifth and sixth classes have 28 students. How many students does Monica see each day?

New Question: Monica is a teacher with 6 classes per day. Her first class has 20 students, her second and third classes have 25 students, and her fourth class has 10 students. Her fifth and sixth classes have 28 students. How many students does Monica see each day in all of her classes?

Original Question: Emily went to the store and bought art supplies for $20 and 2 skirts that cost the same amount of money. She spent a total of $50. How much did Emily pay for each of the skirts?

New Question: Emily went to the store and bought art supplies for $20 and 2 skirts for a total of $50. How much did Emily pay for each of the skirts?

Original Question: John’s neighbor tells him to walk his dog for 1 hour each day for a total of $10. He does this for April, save for the 4 Sundays in April. He later spent $50 on books and gave his sister Kaylee the same amount. How much money did John have left?

New Question: John’s neighbor tells him to walk his dog for April (30 days excluding 4 Sundays) for a total of $10 each day. He later spent $50 on books and gave his sister Kaylee the same amount. How much money did John have left after these expenses?

Original Question: Three years ago, Bethany was twice the age of her younger sister. In 5 years, her younger sister will be 16. How old is Bethany now?

New Question: Three years ago, Bethany was twice the age of her younger sister, who is currently 11 years old. How old is Bethany now?

Original Question: At the bookstore, Sarah bought 6 paperback books and 4 hardback books. Her brother bought one-third as many paperback books as Sarah bought, and two times the number of hardback books that she bought. How many books did her brother buy in total?

New Question: At the bookstore, Sarah bought 6 paperback books and 4 hardback books. Her brother bought 2 paperback books and 8 hardback books. How many books did her brother buy in total?

Original Question: Sandra had 2 different bags of candy. Each of her bags had 6 pieces of candy left. Her brother, Roger, also had 2 bags of candy. One of his bags of candy had 11 pieces left and the other had 3 pieces left. How much more candy did Roger have?

New Question: Sandra had 2 bags of candy, each with 6 pieces left. Her brother, Roger, had 2 bags of candy, one with 11 pieces left and the other with 3 pieces left. How many more pieces of candy did Roger have than Sandra?

Original Question: Joan wants to visit her family who live 480 miles away. If she drives at a rate of 60 mph and takes a lunch break taking 30 minutes, and 2 bathroom breaks taking 15 minutes each, how many hours did it take her to get there?

New Question: Joan wants to visit her family who live 480 miles away. If she drives at a rate of 60 mph and takes a lunch break of 30 minutes, and 2 bathroom breaks of 15 minutes each, how many hours(60 minutes = 1 hour) does it take her to get there?

Original Question: James gets a fleet of gas transportation vans. He gets 6 vans. 2 of them are 8000 gallons. 1 of them is 30% less than that. The remaining trucks are 50% larger than the 2 trucks. How many gallons can he transport?

New Question: James has acquired a fleet of gas transportation vans. He has 6 vans in total. 2 of the vans have a capacity of 8000 gallons, while the other van has a capacity of 5600 gallons (30% less than the first two vans). The remaining 3 vans have a capacity of 12000 gallons (50% larger than the first two vans). What is the total capacity of the fleet in gallons?

Please rewrite new versions of the original mathematical question (including the context and the final question) to be more understandable and easy to answer. Don’t omit any useful information, especially the numbers.

Original Question: Angelo and Melanie want to plan how many hours over the next week they should study together for their test next week. They have 2 chapters of their textbook to study and 4 worksheets to memorize. They figure out that they should dedicate 3 hours to each chapter of their textbook and 1.5 hours for each worksheet. If they plan to study no more than 4 hours each day, how many days should they plan to study total over the next week if they take a 10-minute break every hour, include 3 10-minute snack breaks each day, and 30 minutes for lunch each day?

New Question: Angelo and Melanie want to plan how many hours over the next week they should study together for their test next week. They have 2 chapters of their textbook to study and they decide to dedicate 3 hours to each chapter. They also have 4 worksheets to memorize, and they decide to dedicate 1.5 hours for each worksheet. Taking into account 10-minute breaks every hour, if they plan to study no more than 4 hours each day including 3 10-minute snack breaks each day, and 30 minutes for lunch each day, how many days should they plan to study total over the next week?

Original Question: Mark’s basketball team scores 25 2 pointers, 8 3 pointers and 10 free throws. Their opponents score double the 2 pointers but half the 3 pointers and free throws. What’s the total number of points scored by both teams added together?

New Question: Mark’s basketball team scores 25 2 pointers, 8 3 pointers and 10 free throws. Their opponents score 50 2 pointers, 4 3 pointers and 5 free throws. Both teams score 75 2 pointers, 12 3 pointers and 15 free throws. What is the total number of points scored by both teams combined?

Original Question: Bella has two times as many marbles as frisbees. She also has 20 more frisbees than deck cards. If she buys 2/5 times more of each item, what would be the total number of the items she will have if she currently has 60 marbles?

New Question: Bella currently has 60 marbles, and she has twice as many marbles as frisbees and 20 more frisbees than deck cards. She buys 2/5 times more of each item. What would be the total number of the items she will have?

Original Question: A group of 4 fruit baskets contains 9 apples, 15 oranges, and 14 bananas in the first three baskets and 2 less of each fruit in the fourth basket. How many fruits are there?

New Question: There is a group of 4 fruit baskets. The first three baskets each contains 9 apples, 15 oranges, and 14 bananas, and 7 apples, 13 oranges, and 12 bananas in the fourth basket. How many fruits are there in total?

Original Question: You can buy 4 apples or 1 watermelon for the same price. You bought 36 fruits evenly split between oranges, apples and watermelons, and the price of 1 orange is $0.50. How much does 1 apple cost if your total bill was $66?

New Question: You bought 36 fruits, with an equal number of oranges, apples and watermelons. The price of 1 watermelon equals to 4 apples, and the price of 1 orange is $0.50. If your total bill was $66, how much does 1 apple cost?

Original Question: Susy goes to a large school with 800 students, while Sarah goes to a smaller school with only 300 students. At the start of the school year, Susy had 100 social media followers. She gained 40 new followers in the first week of the school year, half that in the second week, and half of that in the third week. Sarah only had 50 social media followers at the start of the year, but she gained 90 new followers the first week, a third of that in the second week, and a third of that in the third week. After three weeks, how many social media followers did the girl with the most total followers have?

New Question: At the start of the school year, Susy had 100 social media followers and Sarah had 50 social media followers. Susy gained 40 followers in the first week, 20 in the second week, and 10 in the third week. Sarah gained 90 followers in the first week, 30 in the second week, and 10 in the third week. After three weeks, how many social media followers did the girl with the most total followers have?

Original Question: Sam bought a dozen boxes, each with 30 highlighter pens inside, for $10 each box. He rearranged five of these boxes into packages of six highlighters each and sold them for $3 per package. He sold the rest of the highlighters separately at the rate of three pens for $2. How much profit did he make in total, in dollars?

New Question: Sam bought 12 boxes for $10 each, and each contains 30 highlighter pens. 1 package contains 6 highlighters. He rearranged five of these boxes into packages and sold them for $3 per package. He sold the remaining highlighters separately at the price of $2 for every three one. How much profit did Sam make in total, in dollars?

Original Question: In a certain school, 2/3 of the male students like to play basketball, but only 1/5 of the female students like to play basketball. What percent of the population of the school do not like to play basketball if the ratio of the male to female students is 3:2 and there are 1000 students?

New Question: In a certain school, there is a total of 1000 students, while 3 male students for every 2 female students. So there are 600 male students and 2/3 of the male students like to play basketball, and there are 400 female students but only 1/5 of the female students like to play basketball. What percent of the population of the school do not like to play basketball?

Enhance Reasoning in Large Language Models via Problem Refinement (2024)

References

Top Articles
Latest Posts
Article information

Author: Van Hayes

Last Updated:

Views: 6296

Rating: 4.6 / 5 (46 voted)

Reviews: 93% of readers found this page helpful

Author information

Name: Van Hayes

Birthday: 1994-06-07

Address: 2004 Kling Rapid, New Destiny, MT 64658-2367

Phone: +512425013758

Job: National Farming Director

Hobby: Reading, Polo, Genealogy, amateur radio, Scouting, Stand-up comedy, Cryptography

Introduction: My name is Van Hayes, I am a thankful, friendly, smiling, calm, powerful, fine, enthusiastic person who loves writing and wants to share my knowledge and understanding with you.