An Interpretable Client Decision Tree Aggregation process for Federated Learning
Abstract: Trustworthy Artificial Intelligence solutions are essential in today's data-driven applications, prioritizing principles such as robustness, safety, transparency, explainability, and privacy among others. This has led to the emergence of Federated Learning as a solution for privacy and distributed machine learning. While decision trees, as self-explanatory models, are ideal for collaborative model training across multiple devices in resource-constrained environments such as federated learning environments for injecting interpretability in these models. Decision tree structure makes the aggregation in a federated learning environment not trivial. They require techniques that can merge their decision paths without introducing bias or overfitting while keeping the aggregated decision trees robust and generalizable. In this paper, we propose an Interpretable Client Decision Tree Aggregation process for Federated Learning scenarios that keeps the interpretability and the precision of the base decision trees used for the aggregation. This model is based on aggregating multiple decision paths of the decision trees and can be used on different decision tree types, such as ID3 and CART. We carry out the experiments within four datasets, and the analysis shows that the tree built with the model improves the local models, and outperforms the state-of-the-art.
- Trustworthy artificial intelligence, Electronic Markets 31 (2020).
- Connecting the dots in trustworthy artificial intelligence: From AI principles, ethics, and key requirements to responsible AI systems and regulation, Information Fusion 99 (2023) 101896.
- Trustworthy AI: From principles to practices, ACM Comput. Surv. 55 (2023).
- Explainability of artificial intelligence methods, applications and challenges: A comprehensive survey, Information Sciences 615 (2022) 238–292.
- Federated learning in smart cities: Privacy and security survey, Information Sciences 632 (2023) 833–857.
- Towards risk-free trustworthy artificial intelligence: Significance and requirements, International Journal of Intelligent Systems 2023 (2023) 41.
- A hybrid approach to privacy-preserving federated learning, in: Proceedings of the 12th ACM Workshop on Artificial Intelligence and Security, Association for Computing Machinery, 2019, p. 1–11.
- Federated forest, IEEE Transactions on Big Data 8 (2022) 843–854.
- Federated Random Forests can improve local performance of predictive models for various healthcare applications, Bioinformatics 38 (2022) 2278–2286.
- Bofrf: A novel boosting-based federated random forest algorithm on horizontally partitioned data, IEEE Access 10 (2022) 89835–89851.
- J. R. Quinlan, Induction of decision trees, Machine Learning 1 (1986) 81–106.
- Federated learning: Strategies for improving communication efficiency, in: NIPS Workshop on Private Multi-Party Machine Learning, 2016.
- Towards federated learning: An overview of methods and applications, Wiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery 13 (2023) e1486.
- A tutorial on federated learning from theory to practice: Foundations, software frameworks, exemplary use cases, and selected trends, IEEE/CAA Journal of Automatica Sinica 11 (2024) 1–27.
- Building trusted federated learning: Key technologies and challenges, Journal of Sensor and Actuator Networks 12 (2023) 1–18.
- Fedcs: Efficient communication scheduling in decentralized federated learning, Information Fusion (2023) 102028.
- L. Breiman, Random forests, Machine learning 45 (2001) 5–32.
- LightGBM: A highly efficient gradient boosting decision tree, in: Advances in Neural Information Processing Systems, volume 30, Curran Associates, Inc., 2017.
- T. Chen, C. Guestrin, Xgboost: A scalable tree boosting system, in: Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD ’16, ACM, 2016.
- A practical tutorial on bagging and boosting based ensembles for machine learning: Algorithms, software tools, performance study, practical perspectives and opportunities, Information Fusion 64 (2020) 205–237.
- Explainable artificial intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI, Information Fusion 58 (2020) 82–115.
- Federated learning and differential privacy: Software tools analysis, the sherpa.ai FL framework and methodological guidelines for preserving data privacy, Information Fusion 64 (2020) 270–292.
- Revfrf: Enabling cross-domain random forest training with revocable federated learning, IEEE Transactions on Dependable and Secure Computing 19 (2022) 3671–3685.
- eFL-Boost: Efficient federated learning for gradient boosting decision trees, IEEE Access 10 (2022) 43954–43963.
- OpBoost: a vertical federated tree boosting framework based on order-preserving desensitization, Proc. VLDB Endow. 16 (2022) 202–215.
- SecureBoost: A lossless federated learning framework, IEEE Intelligent Systems 36 (2021) 87–98.
- Secureboost+: A high performance gradient boosting tree framework for large scale vertical federated learning, arXiv preprint arXiv:2110.10927 (2021).
- Practical federated gradient boosting decision trees, in: Proceedings of the AAAI conference on artificial intelligence, volume 34, 2020, pp. 4642–4649.
- O. Sagi, L. Rokach, Explainable decision forest: Transforming a decision forest into an interpretable tree, Information Fusion 61 (2020) 124 – 138.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.