Research Papers Summarized

Looking forward to read research papers and sharing what I understand

Follow publication

Photo by Markus Winkler on Unsplash

E6 : Least-to-most Prompting

Praveen Thenraj
Research Papers Summarized
3 min readJul 16, 2023

Breaking a complex problem into simpler subproblems and solving them so as to solve the complex problem

Paper Name : Least-to-most prompting enables complex reasoning in large language models

Paper URL : https://arxiv.org/abs/2205.10625

Authors : Google Research, Brain Team - Denny Zhou, Nathanael Schärli, Le Hou, Jason Wei, Nathan Scales, Xuezhi Wang, Dale Schuurmans, Claire Cui, Olivier Bousquet, Quoc Le, Ed Chi

Conference : ICLR 2023

Please find the annotated paper here

Problem Statement :

  1. CoT tends to perform poorly on problems that are harder than the exemplars in the prompt
  2. Prompting has always been unidirectional - a tool to instruct LLM rather than being bidirectional which allows taking feedbacks from LLM as well

Solution :

  • Solution consists of two stages namely decomposition and subproblem solving
  • Decomposition - Breaking down a complex problem into multiple subproblems using few shot prompting.
  • Subproblem solving - Solving the subproblems using few shot prompting to generate the result. Passing the solved subproblem and its result as part of prompt to solve the next subproblem.

Experimentation :

  • The approach was tested on three major tasks - symbolic manipulation (last letter concatenation), compositional generalisation (SCAN), maths reasoning (GSM8K)
  • As part of the experimentation, least-to-most prompting was used along with text-davinci and code-davinci models to enable reasoning in LLMs.
  • Few shot prompting, CoT and least-to-most prompting techniques were compared against each other on the three tasks mentioned.

Observations :

  • As part of symbolic manipulation, last letter concatenation task was considered. A list with different number of words (L=4,6,8,10,12) were considered.
  • Least-to-most prompting outperformed both CoT and few-shot learning under all scenarios. It outperformed CoT by a huge margin especially for long length lists (L=12).
  • Least-to-most prompting achieved an accuracy of 99.7% on compositional generalisation task with code-davinci-002 which was way ahead of the result achieved by CoT (16.7%).
  • The accuracy achieved in compositional generalisation task was ahead of all the existing prior works without any neural network training or fine-tuning.
  • In maths reasoning, least-to-most prompting outperforms CoT by a narrow margin of 2% only.
  • But analysis show that there was a significant improvement of 6% over CoT when solving complex problems that requires decomposing into 5 or more steps. This emphasises the need for solving a complex problem into smaller subproblems.

Limitations :

  • Decomposition prompts fail to generalise to a different domain and sometime within domain problems.
  • It was observed that decomposition generalisation was difficult even for different problems within the GSM8K dataset itself thus reducing the performance gains of least-to-most prompting.

Conclusion :

  • Least-to-most prompting opens way to mimicing human approaches to solve complex problems.
  • Least-to-most prompting can be integrated along with CoT to make it even more powerful for reasoning in LLMs.
  • Least-to-most prompting can be that step towards making prompts being bidirectional which involves instructing a LLM and as well getting its feedback.

Sign up to discover human stories that deepen your understanding of the world.

Free

Distraction-free reading. No ads.

Organize your knowledge with lists and highlights.

Tell your story. Find your audience.

Membership

Read member-only stories

Support writers you read most

Earn money for your writing

Listen to audio narrations

Read offline with the Medium app

Research Papers Summarized
Research Papers Summarized

Published in Research Papers Summarized

Looking forward to read research papers and sharing what I understand

Praveen Thenraj
Praveen Thenraj

Written by Praveen Thenraj

A Machine Learning Engineer with interest in NLP

No responses yet

Write a response