AI establishments develop heterogeneous fashions for particular duties however face information shortage challenges throughout coaching. Conventional Federated Studying (FL) helps solely homogeneous mannequin collaboration, which wants an identical architectures throughout all shoppers. Nonetheless, shoppers develop mannequin architectures for his or her distinctive necessities. Furthermore, sharing effort-intensive domestically educated fashions incorporates mental property and reduces individuals’ curiosity in participating in collaborations. Heterogeneous Federated Studying (HtFL) addresses these limitations, however the literature lacks a unified benchmark for evaluating HtFL throughout numerous domains and features.
Background and Classes of HtFL Strategies
Current FL benchmarks concentrate on information heterogeneity utilizing homogeneous shopper fashions however neglect actual eventualities that contain mannequin heterogeneity. Consultant HtFL strategies fall into three predominant classes addressing these limitations. Partial parameter sharing strategies comparable to LG-FedAvg, FedGen, and FedGH keep heterogeneous characteristic extractors whereas assuming homogeneous classifier heads for information switch. Mutual distillation, comparable to FML, FedKD, and FedMRL, trains and shares small auxiliary fashions by distillation strategies. Prototype sharing strategies switch light-weight class-wise prototypes as international information, gathering native prototypes from shoppers, and gathering them on servers to information native coaching. Nonetheless, it stays unclear whether or not present HtFL strategies carry out persistently throughout numerous eventualities.
Introducing HtFLlib: A Unified Benchmark
Researchers from Shanghai Jiao Tong College, Beihang College, Chongqing College, Tongji College, Hong Kong Polytechnic College, and The Queen’s College of Belfast have proposed the primary Heterogeneous Federated Studying Library (HtFLlib), a simple and extensible methodology for integrating a number of datasets and mannequin heterogeneity eventualities. This methodology integrates:
- 12 datasets throughout numerous domains, modalities, and information heterogeneity eventualities
- 40 mannequin architectures starting from small to giant, throughout three modalities.
- A modularized and easy-to-extend HtFL codebase with implementations of 10 consultant HtFL strategies.
- Systematic evaluations protecting accuracy, convergence, computation prices, and communication prices.
Datasets and Modalities in HtFLlib
HtFLlib incorporates detailed information heterogeneity eventualities divided into three settings: Label Skew with Pathological and Dirichlet as subsettings, Characteristic Shift, and Actual-World. It integrates 12 datasets, together with Cifar10, Cifar100, Flowers102, Tiny-ImageNet, KVASIR, COVIDx, DomainNet, Camelyon17, AG Information, Shakespeare, HAR, and PAMAP2. These datasets differ considerably in area, information quantity, and sophistication numbers, demonstrating HtFLlib’s complete and versatile nature. Furthermore, researchers’ predominant focus is on picture information, particularly the label skew setting, as picture duties are essentially the most generally used duties throughout numerous fields. The HtFL strategies are evaluated throughout picture, textual content, and sensor sign duties to judge their respective strengths and weaknesses.
Efficiency Evaluation: Picture Modality
For picture information, most HtFL strategies present decreased accuracy as mannequin heterogeneity will increase. The FedMRL exhibits superior energy by its mixture of auxiliary international and native fashions. When introducing heterogeneous classifiers that make partial parameter sharing strategies inapplicable, FedTGP maintains superiority throughout numerous settings as a consequence of its adaptive prototype refinement skill. Medical dataset experiments with black-boxed pre-trained heterogeneous fashions reveal that HtFL enhances mannequin high quality in comparison with pre-trained fashions and achieves higher enhancements than auxiliary fashions, comparable to FML. For textual content information, FedMRL’s benefits in label skew settings diminish in real-world settings, whereas FedProto and FedTGP carry out comparatively poorly in comparison with picture duties.
Conclusion
In conclusion, researchers launched HtFLlib, a framework that addresses the crucial hole in HtFL benchmarking by offering unified analysis requirements throughout numerous domains and eventualities. HtFLlib’s modular design and extensible structure present an in depth benchmark for each analysis and sensible purposes in HtFL. Furthermore, its skill to help heterogeneous fashions in collaborative studying opens the best way for future analysis into using advanced pre-trained giant fashions, black-box programs, and different architectures throughout totally different duties and modalities.
Take a look at the Paper and GitHub Page. All credit score for this analysis goes to the researchers of this venture. Additionally, be at liberty to comply with us on Twitter and don’t neglect to hitch our 100k+ ML SubReddit and Subscribe to our Newsletter.

Sajjad Ansari is a last yr undergraduate from IIT Kharagpur. As a Tech fanatic, he delves into the sensible purposes of AI with a concentrate on understanding the impression of AI applied sciences and their real-world implications. He goals to articulate advanced AI ideas in a transparent and accessible method.