TY - JOUR
T1 - Mix2SFL
T2 - Two-Way Mixup for Scalable, Accurate, and Communication-Efficient Split Federated Learning
AU - Oh, Seungeun
AU - Nam, Hyelin
AU - Park, Jihong
AU - Vepakomma, Praneeth
AU - Raskar, Ramesh
AU - Bennis, Mehdi
AU - Kim, Seong Lyun
N1 - Publisher Copyright:
© 2015 IEEE.
PY - 2024/6/1
Y1 - 2024/6/1
N2 - In recent years, split learning (SL) has emerged as a promising distributed learning framework that can utilize Big Data in parallel without privacy leakage while reducing client-side computing resources. In the initial implementation of SL, however, the server serves multiple clients sequentially incurring high latency. Parallel implementation of SL can alleviate this latency problem, but existing Parallel SL algorithms compromise scalability due to its fundamental structural problem. To this end, our previous works have proposed two scalable Parallel SL algorithms, dubbed SGLR and LocFedMix-SL, by solving the aforementioned fundamental problem of the Parallel SL structure. In this article, we propose a novel Parallel SL framework, coined Mix2SFL, that can ameliorate both accuracy and communication-efficiency while still ensuring scalability. Mix2SFL first supplies more samples to the server through a manifold mixup between the smashed data uploaded to the server as in SmashMix of LocFedMix-SL, and then averages the split-layer gradient as in GradMix of SGLR, followed by local model aggregation as in SFL. Numerical evaluation corroborates that Mix2SFL achieves improved performance in both accuracy and latency compared to the state-of-the-art SL algorithm with scalability guarantees. Moreover, its convergence speed as well as privacy guarantee are validated through the experimental results.
AB - In recent years, split learning (SL) has emerged as a promising distributed learning framework that can utilize Big Data in parallel without privacy leakage while reducing client-side computing resources. In the initial implementation of SL, however, the server serves multiple clients sequentially incurring high latency. Parallel implementation of SL can alleviate this latency problem, but existing Parallel SL algorithms compromise scalability due to its fundamental structural problem. To this end, our previous works have proposed two scalable Parallel SL algorithms, dubbed SGLR and LocFedMix-SL, by solving the aforementioned fundamental problem of the Parallel SL structure. In this article, we propose a novel Parallel SL framework, coined Mix2SFL, that can ameliorate both accuracy and communication-efficiency while still ensuring scalability. Mix2SFL first supplies more samples to the server through a manifold mixup between the smashed data uploaded to the server as in SmashMix of LocFedMix-SL, and then averages the split-layer gradient as in GradMix of SGLR, followed by local model aggregation as in SFL. Numerical evaluation corroborates that Mix2SFL achieves improved performance in both accuracy and latency compared to the state-of-the-art SL algorithm with scalability guarantees. Moreover, its convergence speed as well as privacy guarantee are validated through the experimental results.
KW - Accuracy
KW - communication efficiency
KW - distributed machine learning
KW - federated learning
KW - privacy
KW - scalability
KW - split learning
UR - http://www.scopus.com/inward/record.url?scp=85181826253&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85181826253&partnerID=8YFLogxK
U2 - 10.1109/TBDATA.2023.3328424
DO - 10.1109/TBDATA.2023.3328424
M3 - Article
AN - SCOPUS:85181826253
SN - 2332-7790
VL - 10
SP - 238
EP - 248
JO - IEEE Transactions on Big Data
JF - IEEE Transactions on Big Data
IS - 3
ER -