ISSN: 2182-2069 (printed) / ISSN: 2182-2077 (online)
FedProx: FedSplit Algorithm based Federated Learning for Statistical and System Heterogeneity in Medical Data Communication
Distributed machine learning offers more practical and efficient use cases than conventional centralized machine learning. Nevertheless, not all security needs can be satisfied by distributed learning. In medical industry, more and more individuals are adopting Internet of Things (IoT) devices to capture their personal data for medical diagnosis and treatment. Through the use of federated learning, it is possible to secure user data while simultaneously training on massive amounts of dispersed data. The heterogeneity of client data is a well-known difficulty in federated learning (FL) contexts. The solution that developed was customized federated learning (PFL), a framework for developing local models for clients' requirements. It is common practice in PFL to construct two models at once, one for local usage and one for global use; the first is used for generalization, while the second is used to inform and update the first one. To build better customized models, it is vital to realize that both global and local models may be enhanced to increase their generalization potential. Secure heterogeneity medical data collection and training has emerged as the top priority in FL. In order to address statistical and system heterogeneity, this work presents a novel hybrid federated learning approach that uses FedProx: FedSplit Algorithm. The presence and severity of data heterogeneity determine the kind of federated learning approaches that could be necessary. The FedProx method averages the changes to the local model if the user data is horizontally partitioned, which means that different samples have the same features. Techniques such as FedSplit may be required to align the feature spaces or separate the model layers when working with data sources that are vertically partitioned, indicating they contain unique features but overlap samples. As a result of statistical variability, learning across data from different distributions is challenging, and device-level systems limits mean that each device can only do so much work, the FedProx: FedSplit model assures convergence for our method. More specifically, as compared to the present FL model, FedProx: FedSplit shows far more steady and accurate convergence behavior in very diverse conditions, increasing overall test accuracy by an average of 35%.