Large Language Models (LLMs) have successfully catered their way into the challenging areas of Artificial Intelligence. With their amazing ability to produce unique and creative content with great linguistic accuracy and consistency, LLMs are helping out in every industry. Large Language Models are often augmented with reasoning skills and the ability to use different tools. Augmentation basically refers to enhancing or expanding by adding additional elements or features. Augmented LLMs are the ones that are added with external tools and skills in order to increase their performance so that they perform beyond their inherent capabilities.
Applications like Auto-GPT for autonomous task execution have been made possible by Augmented Language Models (ALMs) only. Current ALM attempts mostly rely on the prompting paradigm with interleaved verbal reasoning and tool-calling, which have been effective but also imposes certain limitations. When connecting with external tools, it first necessitates the regular execution and suspension of LLMs, which causes delays and increases token usage. Secondly, LLMs generate tokens based on the previous context, and when halted for tool response, they resume token generation by feeding all historical tokens, which results in significant prompt redundancy leading to high cost in terms of token consumption for commercial LLM services.
To address the challenges, recently, a team of researchers has proposed ReWOO (Reasoning WithOut Observation), a modular paradigm to reduce token consumption. The idea behind ReWOO is to separate the reasoning process of the LLM from external observations, which would help reduce the token consumption significantly. ReWOO minimizes the computational load associated with repeated prompts by separating the reasoning process from external observations.
The key components of an ALM are step-wise reasoning, tool calls, and summarization, which ReWOO divides into three separate modules: Planner, Worker, and Solver. The Planner breaks down a task and formulates a blueprint of interdependent plans, which are each assigned to a Worker. The Worker retrieves external knowledge from tools to provide evidence, and the Solver synthesizes all the plans and evidence to produce the final answer to the initial task to be completed.
To evaluate ReWOO’s performance, the team has carried out a thorough analysis across six open Natural Language Processing (NLP) benchmarks and a curated dataset. The results consistently showed improvements with the proposed methodology, with ReWOO achieving a 5× token efficiency gain and a 4% accuracy improvement on the HotpotQA benchmark, which involves multi-step reasoning tasks. ReWOO also proved to be robust in situations where the external tools had failure issues.
The decoupling of parametric modules from nonparametric tool calls not only increases prompt efficiency but also enables instruction fine-tuning in ReWOO. A 175B parameter GPT3.5 can have its reasoning capability offloaded to a smaller language model, 7B LLaMA, through fine-tuning, leading to a significant reduction in model parameters, which highlights the possibility of developing effective and scalable ALMs.
Consequently, ReWOO is a promising modular paradigm for ALMs as, for the first time, it overcomes the challenges of redundant prompts and computation complexity.
Check Out The Paper and Github link. Don’t forget to join our 22k+ ML SubReddit, Discord Channel, and Email Newsletter, where we share the latest AI research news, cool AI projects, and more. If you have any questions regarding the above article or if we missed anything, feel free to email us at Asif@marktechpost.com
Tanya Malhotra is a final year undergrad from the University of Petroleum & Energy Studies, Dehradun, pursuing BTech in Computer Science Engineering with a specialization in Artificial Intelligence and Machine Learning.
She is a Data Science enthusiast with good analytical and critical thinking, along with an ardent interest in acquiring new skills, leading groups, and managing work in an organized manner.