Home Community Microsoft Researchers Introduce Reprompting: An Iterative Sampling Algorithm that Searches for the Chain-of-Thought (CoT) Recipes for a Given Task without Human Intervention

Microsoft Researchers Introduce Reprompting: An Iterative Sampling Algorithm that Searches for the Chain-of-Thought (CoT) Recipes for a Given Task without Human Intervention

0
Microsoft Researchers Introduce Reprompting: An Iterative Sampling Algorithm that Searches for the Chain-of-Thought (CoT) Recipes for a Given Task without Human Intervention

In recent times, Large Language Models (LLMs) have evolved and transformed Natural Language Processing with their few-shot prompting techniques.  These models have prolonged their usability in almost every domain, starting from Machine translation, Natural Language Understanding, Text completion, sentiment evaluation, speech recognition, and so forth.  With the few-shot prompting approach, LLMs are supplied with just a few examples of a selected task, together with some natural language instructions, and using these; they can adapt and learn tips on how to perform the duty properly.  The tasks requiring iterative steps and constraint propagation include many limitations when using these prompting techniques, to beat which a brand new approach has been introduced.

A team of researchers at Microsoft Research, Redmond, USA, recently introduced a brand new method called Reprompting, which addresses all the restrictions accompanying prompting techniques.  This approach mechanically searches for some useful and effective chain-of-thought (CoT) prompts.  Chain-of-thought prompting helps improve the reasoning ability of enormous language models and helps them perform complex reasoning tasks.  For this, just a few chains of thought demonstrations are provided as exemplars during prompting.  Reprompting finds CoT prompts very efficiently with none human involvement. 

The researchers have used an iterative sampling approach often called Gibbs sampling within the Reprompting algorithm.  It frames the issue as sampling from a joint distribution of CoT recipes.  Because the distribution is difficult to characterize directly, Gibbs Sampling has been used as an approximation method.  This sampling method helps determine the perfect instructions by trying different ones and deciding which works best.

🚀 JOIN the fastest ML Subreddit Community

The Reproompting algorithm begins with a sampling of initial CoT recipes with the assistance of zero-shot prompting, where no prompt information is provided.  Zero-shot prompting enables an LLM to generate task responses without prior training.  The algorithm then iteratively samples latest recipes using previously sampled solutions as parent prompts, and these latest recipes are used to unravel other training problems, aiming to search out a set of prompts that share similar CoT prompts. 

The algorithm has been evaluated on the five Big-Bench Hard (BBH) tasks that require multi-step reasoning.  BBH focuses on tasks which might be believed to be beyond the talents and potentials of the present language models.  ChatGPT and InstructGPT have been used as LLMs for the evaluation of the algorithm.  Upon evaluation, Reprompting has proved to perform higher than the zero-shot, few-shot, and human-written CoT prompting techniques. 

Reprompting also showed significant potential in model combination through the use of different LLMs for initializing and sampling latest recipes.  It might assist in the transfer of information from a stronger model to a weaker model, thus leading to a noticeably higher performance shown by the weaker model.  Reprompting performed higher than the human-written CoT prompting on BBH tasks by as much as 17 points.  The researchers have mentioned that the CoT recipes that work high-quality on one model may not work well on one other, highlighting the necessity for optimizing CoT for every model to have some fairer comparisons.

To sum up, the Reprompting algorithm is an ideal automated approach for locating effective CoT prompts for LLMs without human intervention.  It’s a invaluable approach to addressing the restrictions of existing methods and achieving superior performance on tasks requiring multi-step reasoning.


Take a look at the Paper. Don’t forget to affix our 21k+ ML SubRedditDiscord Channel, and Email Newsletter, where we share the most recent AI research news, cool AI projects, and more. If you may have any questions regarding the above article or if we missed anything, be at liberty to email us at Asif@marktechpost.com

🚀 Check Out 100’s AI Tools in AI Tools Club


Tanya Malhotra is a final 12 months undergrad from the University of Petroleum & Energy Studies, Dehradun, pursuing BTech in Computer Science Engineering with a specialization in Artificial Intelligence and Machine Learning.
She is a Data Science enthusiast with good analytical and important considering, together with an ardent interest in acquiring latest skills, leading groups, and managing work in an organized manner.


➡️ Meet Shiny Data: The World’s #1 Web Data Platform

LEAVE A REPLY

Please enter your comment!
Please enter your name here