Home Community How Do Large Language Models Perform in Long-Form Query Answering? A Deep Dive by Salesforce Researchers into LLM Robustness and Capabilities

How Do Large Language Models Perform in Long-Form Query Answering? A Deep Dive by Salesforce Researchers into LLM Robustness and Capabilities

0
How Do Large Language Models Perform in Long-Form Query Answering? A Deep Dive by Salesforce Researchers into LLM Robustness and Capabilities

While Large Language Models (LLMs) like ChatGPT and GPT-4 have demonstrated higher performance across several benchmarks, open-source projects like MMLU and OpenLLMBoard have quickly progressed in catching up across multiple applications and benchmarks. Understanding their capabilities, constraints, and distinctions becomes more crucial as they enter the brand new era of LLMs with rapid advancements in latest models and methodologies. Although LLMs have demonstrated their ability to generate coherent text in tasks like summarization, more is required about how well they do on LFQA. 

Certainly one of the numerous problems that also must be solved is long-form query answering (LFQA), which has quite a few and significant real-world applications (resembling support forums, troubleshooting, customer support, etc.). Answering such inquiries steadily calls for sophisticated pondering skills to grasp the query and make sense of the fabric that’s dispersed across the unique paper. The important points of the articles are condensed into abstract summaries. They assume that follow-up inquiries from these summaries would necessitate a greater comprehension of the topics connecting various sections of the source material. Moreover, other researchers show that responses that decision for comprehension of greater than a 3rd of a lengthy material are steadily evaluated as “HARD” by people. 

Researchers from Salesforce suggest a scalable assessment approach to check and contrast the differences between huge LLMs and smaller yet successful basic LLMs (resembling Llama-7B, 13B) and their distilled counterparts (resembling Alpaca-7B, 13B). To do that, they indicate that ChatGPT be instructed explicitly to construct complicated questions from document summaries. Their empirical study reveals that follow-up questions created from summaries present a difficult but more realistic setup for assessing the reasoning skills of LLMs on two fronts (complexity of generated questions and response quality of open-source LLMs). They use GPT-4 to find out the response quality on coherence, relevance, factual consistency, and correctness under earlier works because entirely depending on human review for long-form QA is dear and difficult to scale. In addition they do a smaller-scale human evaluation, demonstrating that GPT-4 strongly correlates with human evaluation, making their assessment credible. 

The next are their primary conclusions from this study: 

• They recommend inferring from lengthier contexts by making quite a few runs through the context for > 20% of the time to generate questions from abstractive summaries. 

• Distilled LLMs (Alpaca-7B, 13B) often rely less on context when generating questions from the unique material, but their ability to create questions from document summaries is greatly reduced. 

• For questions derived from summaries (> 16.8%), responses produced by distilled LLMs may be consistent across contexts, but they steadily go off-topic, produce redundant replies, and are only partially accurate. 

• Alpaca-7B and 13B are more sensitive to lengthier contexts (>1024 tokens) than base LLMs (Llama), although they typically produce sensible replies.


Take a look at the Paper. All Credit For This Research Goes To the Researchers on This Project. Also, don’t forget to affix our 30k+ ML SubReddit, 40k+ Facebook Community, Discord Channel, and Email Newsletter, where we share the most recent AI research news, cool AI projects, and more.

For those who like our work, you’ll love our newsletter..


Aneesh Tickoo is a consulting intern at MarktechPost. He’s currently pursuing his undergraduate degree in Data Science and Artificial Intelligence from the Indian Institute of Technology(IIT), Bhilai. He spends most of his time working on projects aimed toward harnessing the facility of machine learning. His research interest is image processing and is captivated with constructing solutions around it. He loves to attach with people and collaborate on interesting projects.


🚀 The top of project management by humans (Sponsored)

LEAVE A REPLY

Please enter your comment!
Please enter your name here