Home Community Meet RAP and LLM Reasoners: Two Frameworks Based on Similar Concepts for Advanced Reasoning with LLMs

Meet RAP and LLM Reasoners: Two Frameworks Based on Similar Concepts for Advanced Reasoning with LLMs

0
Meet RAP and LLM Reasoners: Two Frameworks Based on Similar Concepts for Advanced Reasoning with LLMs

Each passing day brings remarkable progress in Large Language Models (LLMs), resulting in groundbreaking tools and advancements. These LLMs excel in various tasks, including text generation, sentiment classification, text classification, and zero-shot classification. Their capabilities extend beyond these areas, enabling automation of content creation, customer support, and data evaluation, thereby revolutionizing productivity and efficiency.

Recently, the researchers have also began exploring the use and utility of LLMs for reasoning. These models can comprehend complex textual information and draw logical inferences from it. LLMs excel in tasks like question-answering, problem-solving, and decision-making. Nevertheless, LLMs still cannot work like humans combating problems that might be easy for humans, akin to generating motion plans for executing tasks in a given environment or performing complex mathematical, logical, and commonsense reasoning. LLMs struggle with certain tasks because they don’t have an internal world model like humans do. This implies they’ll’t predict how things will likely be in a given situation or simulate long-term outcomes of actions. Humans possess an internal world model, a mental representation of the environment, which enables humans to simulate actions and their effects on the world’s state for deliberate planning during complex tasks.

To beat these issues, researchers have devised a brand new reasoning framework, Reasoning via Planning (RAP). This framework uses a library that permits LLMs to perform complex reasoning using advanced reasoning algorithms. This framework approaches multi-step reasoning methodology as planning and searches for the optimal reasoning chain, which achieves one of the best balance of exploration vs. exploitation with the concept of “World Model” and “Reward.” Other than the RAP paper, the research team also proposes LLM Reasoners. LLM Reasoners is an AI library designed to equip Language Models (LLMs) with the potential to perform intricate reasoning through advanced algorithms. It perceives multi-step reasoning as planning, trying to find probably the most efficient reasoning chain, and optimizing the balance between exploration and exploitation using the concepts of ‘World Model’ and ‘Reward’. All it’s good to do is define a reward function and, optionally, a world model. The LLM Reasoners handle the remainder, encompassing Reasoning Algorithms, Visualization, LLM invocation, and more!

A world model regards the partial solution because the state and easily appends a brand new motion/thought to the state because the state transition. The reward function is crucial in evaluating how well a reasoning step performs. The thought is that a reasoning chain with the next collected reward is more prone to be correct.

The researchers performed extensive research on this framework. They applied RAP to several difficult reasoning problems on mathematical reasoning and logical inference. The sensible results of those tasks show that RAP outperforms several strong baseline methods. When applied to LLaMA33B, RAP surpasses CoT on GPT-4, achieving a formidable 33% relative improvement in plan generation.

Through the reasoning process, the LLM cleverly constructs a reasoning tree by constantly evaluating one of the best possible reasoning steps (actions). To do that, it uses its world model, which is identical LLM used differently. By simulating future outcomes, the LLM estimates potential rewards and uses this information to update its beliefs concerning the current reasoning steps. This fashion, it refines its reasoning by exploring higher alternatives and improving its decisions. This framework offers cutting-edge reasoning algorithms, provides intuitive visualization and Interpretation, and is compatible with some other LLM libraries.

The researchers emphasize that after conducting extensive experiments on various difficult reasoning problems, RAP’s superiority over several contemporary CoT-based reasoning approaches was concluded. The framework even performed higher than advanced GPT-4 in certain settings. The flexibleness of RAP in designing rewards, states, and actions showcases its potential as a flexible framework for tackling various reasoning tasks. It’s fascinating to see how RAP combines planning and reasoning in an progressive way. This approach can potentially revolutionize how we approach LLM reasoning, paving the best way for AI systems to attain human-level strategic pondering and planning.


Try the RAP Paper, LLM Reasoners Project Page, and GitHub. All Credit For This Research Goes To the Researchers on This Project. Also, don’t forget to hitch our 27k+ ML SubReddit, 40k+ Facebook Community, Discord Channel, and Email Newsletter, where we share the newest AI research news, cool AI projects, and more.


Rachit Ranjan is a consulting intern at MarktechPost . He’s currently pursuing his B.Tech from Indian Institute of Technology(IIT) Patna . He’s actively shaping his profession in the sphere of Artificial Intelligence and Data Science and is passionate and dedicated for exploring these fields.


🔥 Gain a competitive
edge with data: Actionable market intelligence for global brands, retailers, analysts, and investors. (Sponsored)

LEAVE A REPLY

Please enter your comment!
Please enter your name here