Introduction
Federated Learning (FL) fundamentally addresses the challenges associated with centralized learning by distributing the training process across multiple clients, enabling parallel processing. This approach also helps to safeguard the privacy of raw data stored on clients by exchanging only the model parameters. However, FL requires local training for each client, which can significantly burden clients with limited battery power and computational resources when dealing with large models such as Deep Learning (DL). Split Learning (SL) has emerged as a solution to mitigate this problem. SL involves breaking down a full DL model into two sub-models that can be trained both on a main server and across distributed clients. This approach alleviates the local training burden associated with FL while preserving data privacy. Nevertheless, SL introduces its own set of challenges, primarily related to the training time overhead, owing to its relay-based training method. In this relay-based approach, only one client trains with the main server at any given time, whereas the other clients remain idle. This sequential training method leads to inefficient distributed processing and a long training latency. To address this challenge, various strategies have been proposed to parallelize the SL training process [1]. Inspired by these efforts, split federated learning, simply called split-fed learning (SFL), has recently been proposed as a novel approach that leverages the strengths of both FL and SL. Unlike SL, in SFL, all clients perform their local training in parallel while actively engaging with the main server and federated server (fed server). In SFL, the fed server plays a pivotal role in aggregating local model updates from clients using predefined aggregation techniques, such as FedAvg. This aggregation process occurred synchronously during each round of training. By introducing this additional aggregation server, SFL seamlessly combines the advantages of both FL and SL [2].