Home Community Microsoft and Columbia Researchers Propose LLM-AUGMENTER: An AI System that Augments a Black-Box LLM with a Set of Plug-and-Play Modules

Microsoft and Columbia Researchers Propose LLM-AUGMENTER: An AI System that Augments a Black-Box LLM with a Set of Plug-and-Play Modules

0
Microsoft and Columbia Researchers Propose LLM-AUGMENTER: An AI System that Augments a Black-Box LLM with a Set of Plug-and-Play Modules

Large language models (LLMs) like GPT-3 are well known for his or her ability to generate coherent and informative natural language texts because of their vast amount of world knowledge. Nevertheless, encoding this data in LLMs is lossy and may result in memory distortion, leading to hallucinations that may be detrimental to mission-critical tasks. Moreover, LLMs cannot encode all crucial information for some applications, making them unsuitable for time-sensitive tasks like news query answering. Although various methods have been proposed to boost LLMs using external knowledge, these typically require fine-tuning LLM parameters, which may be prohibitively expensive. Consequently, there’s a necessity for plug-and-play modules that may be added to a hard and fast LLM to enhance its performance in mission-critical tasks.

The paper proposes a system called LLM-AUGMENTER that addresses the challenges of applying Large Language Models (LLMs) to mission-critical applications. The system is designed to reinforce a black-box LLM with plug-and-play modules to ground its responses in external knowledge stored in task-specific databases. It also includes iterative prompt revision using feedback generated by utility functions to enhance the factuality rating of LLM-generated responses. The system’s effectiveness is validated empirically in task-oriented dialog and open-domain question-answering scenarios, where it significantly reduces hallucinations without sacrificing the fluency and informativeness of reactions. The source code and models of the system are publicly available.

The LLM-Augmenter process involves three primary steps. Firstly, when given a user query, it retrieves evidence from external knowledge sources resembling web searches or task-specific databases. It might also connect the retrieved raw evidence with relevant context and reason on the concatenation to create “evidence chains.” Secondly, the LLM-Augmenter prompts a hard and fast LLM like ChatGPT by utilizing the consolidated evidence to generate a response rooted in evidence. Lastly, LLM-Augmenter checks the generated response and creates a corresponding feedback message. This feedback message modifies and iterates the ChatGPT query until the candidate’s response meets verification requirements.

🚀 Construct high-quality training datasets with Kili Technology and solve NLP machine learning challenges to develop powerful ML applications

The work presented on this study shows that the LLM-Augmenter approach can effectively augment black-box LLMs with external knowledge pertinent to their interactions with users. This augmentation greatly reduces the issue of hallucinations without compromising the fluency and informative quality of the responses generated by the LLMs.

LLM-AUGMENTER’s performance was evaluated on information-seeking dialog tasks using each automatic metrics and human evaluations. Commonly used metrics, resembling Knowledge F1 (KF1) and BLEU-4, were used to evaluate the overlap between the model’s output and the ground-truth human response and the overlap with the knowledge that the human used as a reference during dataset collection. Moreover, the researchers included these metrics that best correlate with human judgment on the DSTC9 and DSTC11 customer support tasks. Other metrics, resembling BLEURT, BERTScore, chrF, and BARTScore, were also considered, as they’re among the many best-performing text generation metrics on the dialog.


Take a look at the Paper and Project. All Credit For This Research Goes To the Researchers on This Project. Also, don’t forget to hitch our 26k+ ML SubRedditDiscord Channel, and Email Newsletter, where we share the newest AI research news, cool AI projects, and more.


Niharika

” data-medium-file=”https://www.marktechpost.com/wp-content/uploads/2023/01/1674480782181-Niharika-Singh-264×300.jpg” data-large-file=”https://www.marktechpost.com/wp-content/uploads/2023/01/1674480782181-Niharika-Singh-902×1024.jpg”>

Niharika is a Technical consulting intern at Marktechpost. She is a 3rd 12 months undergraduate, currently pursuing her B.Tech from Indian Institute of Technology(IIT), Kharagpur. She is a highly enthusiastic individual with a keen interest in Machine learning, Data science and AI and an avid reader of the newest developments in these fields.


🔥 Gain a competitive
edge with data: Actionable market intelligence for global brands, retailers, analysts, and investors. (Sponsored)

LEAVE A REPLY

Please enter your comment!
Please enter your name here