Traditional methods for training vision-language models (VLMs) often require the centralized aggregation of vast datasets, which raises concerns regarding privacy and scalability. Federated learning offers an answer by allowing models to be trained across a distributed network of devices while keeping data locally but adapting VLMs to this framework presents unique challenges.
To handle these challenges, a team of researchers from Intel Corporation and Iowa State University introduced FLORA (Federated Learning with Low-Rank Adaptation) to handle the challenge of coaching vision-language models (VLMs) in federated learning (FL) settings while preserving data privacy and minimizing communication overhead. FLORA fine-tunes VLMs just like the CLIP model by utilizing parameter-efficient adapters, namely Low-Rank Adaptation (LoRA), along side Federated Learning. As an alternative of requiring centralized data mining, FLORA enables model training across decentralized data sources while preserving data privacy and minimizing communication costs. By selectively updating only a small subset of the model’s parameters using LoRA, FLORA accelerates training time and reduces memory usage in comparison with full fine-tuning.
The FLORA method uses LoRA-adapted CLIP models for client-side training and native updates. An Adam optimizer helps with gradient-based optimization. A server then aggregates these updates using a weighted averaging technique much like FedAvg. The Low-Rank Adaptation (LoRA) method is a key a part of FLORA’s success since it adds trainable low-rank matrices to certain layers of a model that has already been trained. This cuts down on the quantity of labor that should be done and the quantity of memory that is required. FLORA improves performance and adapts models more efficiently in federated learning settings by adding LoRA to the CLIP model.
Experimental evaluations reveal FLORA’s effectiveness across various datasets and learning environments. FLORA consistently outperforms traditional FL methods in each IID and non-IID settings, demonstrating superior accuracy and flexibility. Also, FLORA’s efficiency evaluation shows that it uses much less memory and communication in comparison with baseline methods, which shows that it may very well be utilized in real-world federated learning situations. Just a few-shot evaluation further confirms FLORA’s proficiency in managing data scarcity and distribution variability, showcasing its robust performance even with limited training examples.
In conclusion, FLORA presents a promising solution to the challenge of coaching vision-language models in federated learning settings. By leveraging Federated Learning and Low-Rank Adaptation, FLORA enables efficient model adaptation while preserving data privacy and minimizing communication overhead. The methodology’s performance across various datasets and learning environments underscores its potential to revolutionize federated learning for VLMs. The superior accuracy, efficiency, and flexibility that FLORA can achieve makes it a powerful solution for coping with the difficulties of real-world data challenges in distributed learning environments.
Take a look at the Paper. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter. Join our Telegram Channel, Discord Channel, and LinkedIn Group.
If you happen to like our work, you’ll love our newsletter..
Don’t Forget to affix our 40k+ ML SubReddit
Pragati Jhunjhunwala is a consulting intern at MarktechPost. She is currently pursuing her B.Tech from the Indian Institute of Technology(IIT), Kharagpur. She is a tech enthusiast and has a keen interest within the scope of software and data science applications. She is all the time reading concerning the developments in numerous field of AI and ML.