IJMECS Vol. 17, No. 2, 8 Apr. 2025
Cover page and Table of Contents: PDF (size: 1027KB)
PDF (1027KB), PP.87-110
Views: 0 Downloads: 0
Feature Extraction, Fraud Review, Machine Learning, Teacher Performance, Instructor Performance
Federated Learning (FL) is an emerging machine learning approach with promising applications. In this paper, FL is comprehensively examined in relation to teacher performance evaluation. Through FL, teachers can be evaluated based on data-driven metrics while preserving data privacy. There are several benefits, including data privacy preservation, collaborative learning, scalability, and privacy-preserving insights. Additionally, it faces problems related to communication efficiency, system heterogeneity, and statistical heterogeneity. To address these issues, we propose a novel clustering-based technique in federated learning. The technique aims to overcome the challenges of system heterogeneity and improve communication efficiency. We provide a comprehensive review of existing research on clustering techniques in the context of federated learning, offering insights into the state of the art in this field. In addition, we emphasize the need for advanced compression methods, enhanced privacy-preserving mechanisms, and robust aggregation algorithms for future federated learning research. To address these challenges, we present a clustering-based approach to address the merits and challenges of federated learning The clustering-based approach we propose in this research demonstrates promising results in terms of reducing communication overhead and improving model convergence in federated learning. These findings suggest that incorporating clustering techniques can significantly enhance the efficiency and effectiveness of federated learning algorithms, paving the way for more scalable and privacy-preserving distributed machine learning systems. The findings of this study suggest that clustering techniques can improve the efficiency and scalability of federated learning.
Ariful Islam, Debajyoti Karmaker, Abhijit Bhowmik, Md Masum Billah, Md Iftekharul Mobin, Noorhuzaimi Mohd Noor, "Unlocking Educational Excellence: Leveraging Federated Learning for Enhanced Instructor Evaluation and Student Success", International Journal of Modern Education and Computer Science(IJMECS), Vol.17, No.2, pp. 87-110, 2025. DOI:10.5815/ijmecs.2025.02.04
[1]S. Shen, T. Zhu, D. Wu, W. Wang, and W. Zhou, “From Distributed Machine Learning To Federated Learning: In The View Of Data Privacy And Security,” Concurrency and Computation, vol. 34, no. 16, Jul. 2022, doi: 10.1002/cpe.6002.
[2]C. Briggs, Z. Fan, and P. Andras, “Federated learning with hierarchical clustering of local updates to improve training on non-iid data,” arXiv preprint arXiv:2004.11791, 2020
[3]T. Li, A. K. Sahu, A. Talwalkar, and V. Smith, “Federated Learning: Challenges, Methods, and Future Directions,” IEEE Signal Process. Mag., vol. 37, no. 3, pp. 50–60, May 2020, doi: 10.1109/MSP.2020.2975749.
[4]K. Bonawitz et al., “Towards Federated Learning at Scale: System Design.” arXiv, Mar. 22, 2019. Accessed: Mar. 20, 2023. [Online]. Available: http://arxiv.org/abs/1902.01046
[5]L. Li, Y. Fan, M. Tse, and K.-Y. Lin, “A review of applications in federated learning,” Computers & Industrial Engineering, vol. 149, p. 106854, Nov. 2020, doi: 10.1016/j.cie.2020.106854.
[6]A. Reisizadeh, A. Mokhtari, H. Hassani, A. Jadbabaie, and R. Pedarsani, “Fedpaq: A communication-efficient federated learning method with periodic averaging and quantization,” in International Conference on Artificial Intelligence and Statistics. PMLR, 2020, pp. 2021–2031
[7]“Challenges and future directions of secure federated learning: a survey | SpringerLink.” https://link.springer.com/article/10.1007/s11704-021-0598-z (accessed Mar. 21, 2023).
[8]C. Ma, J. Li, M. Ding, K. Wei, W. Chen, and H. V. Poor, “Federated Learning with Unreliable Clients: Performance Analysis and Mechanism Design.” arXiv, Jul. 31, 2021. doi: 10.48550/arXiv.2105.06256.
[9]Latif, U., Khan., Madyan, Alsenwi., Zhu, Han., Choong, Seon, Hong. (2020). Self Organizing Federated Learning Over Wireless Networks: A Socially Aware Clustering Approach. 453-458. doi: 10.1109/ICOIN48656.2020.9016505
[10] T. Li, A. K. Sahu, M. Zaheer, M. Sanjabi, A. Talwalkar, and V. Smith, “Federated Optimization in Heterogeneous Networks,” 2018, [Online]. Available: http://arxiv.org/abs/1812.06127.
[11]Y. Zhao, M. Li, L. Lai, N. Suda, D. Civin, and V. Chandra, “Federated learning with non-iid data,” arXiv preprint arXiv:1806.00582, 2018
[12]C. Xu, Y. Qu, Y. Xiang, and L. Gao, “Asynchronous Federated Learning on Heterogeneous Devices: A Survey.” arXiv, Mar. 30, 2023. Accessed: Jul. 21, 2023. [Online]. Available: http://arxiv.org/abs/2109.04269
[13]Hanlin Tang, Shaoduo Gan, Ce Zhang, Tong Zhang, and Ji Liu. 2018. Communication compression for decentralized training. In Proceedings of the 32nd International Conference on Neural Information Processing Systems (NIPS'18). Curran Associates Inc., Red Hook, NY, USA, 7663–7673.
[14]M. Asad, A. Moustafa, and T. Ito, “FedOpt: Towards Communication Efficiency and Privacy Preservation in Federated Learning,” Applied Sciences, vol. 10, no. 8, Art. no. 8, Jan. 2020, doi: 10.3390/app10082864.
[15]M. Chen, N. Shlezinger, H. V. Poor, Y. C. Eldar, and S. Cui, “Communication-efficient federated learning,” Proceedings of the National Academy of Sciences, vol. 118, no. 17, p. e2024789118, 2021, doi: 10.1073/pnas.2024789118.
[16]J. Konečný, H. B. McMahan, F. X. Yu, P. Richtárik, A. T. Suresh, and D. Bacon, “Federated Learning: Strategies for Improving Communication Efficiency.” arXiv, Oct. 30, 2017. Accessed: May 10, 2023. [Online]. Available: http://arxiv.org/abs/1610.05492
[17]Kang, D.; Ahn, C.W. Communication Cost Reduction with Partial Structure in Federated Learning. Electronics 2021, 10, 2081. https://doi.org/10.3390/ electronics10172081
[18]G. Yang, K. Mu, C. Song, Z. Yang, and T. Gong, “RingFed: Reducing Communication Costs in Federated Learning on Non-IID Data.” arXiv, Jul. 19, 2021. Accessed: May 10, 2023. [Online]. Available: http://arxiv.org/abs/2107.08873
[19]Park, S.; Suh, Y.; Lee, J. FedPSO: Federated Learning Using Particle Swarm Optimization to Reduce Communication Costs. Sensors 2021, 21, 600. https:// doi.org/10.3390/s21020600
[20]A. K. Abasi, M. Aloqaily, M. Guizani and F. Karray, "Sine Cosine Algorithm for Reducing Communication Costs of Federated Learning," 2022 IEEE International Mediterranean Conference on Communications and Networking (MeditCom), Athens, Greece, 2022, pp. 55-60, doi: 10.1109/MeditCom55741.2022.9928614.
[21]S. Wang et al., “Adaptive Federated Learning in Resource Constrained Edge Computing Systems.” arXiv, Feb. 16, 2019. Accessed: May 10, 2023. [Online]. Available: http://arxiv.org/abs/1804.05271
[22]Y. Zhao, M. Li, L. Lai, N. Suda, D. Civin, and V. Chandra, “Federated Learning with Non-IID Data,” 2018, doi: 10.48550/arXiv.1806.00582.
[23]Hou, C., Thekumparampil, K.K., Fanti, G.C., & Oh, S. (2021). Reducing the Communication Cost of Federated Learning through Multistage Optimization. ArXiv, abs/2108.06869.
[24]X. Yao, C. Huang, and L. Sun, “Two-Stream Federated Learning: Reduce the Communication Costs,” in 2018 IEEE Visual Communications and Image Processing (VCIP), Taichung, Taiwan: IEEE, Dec. 2018, pp. 1–4. doi: 10.1109/VCIP.2018.8698609
[25]X. Yao, T. Huang, C. Wu, R.-X. Zhang, and L. Sun, “Federated Learning with Additional Mechanisms on Clients to Reduce Communication Costs.” arXiv, Sep. 01, 2019. Accessed: May 10, 2023. [Online]. Available: http://arxiv.org/abs/1908.05891
[26]Mitra, A., Jaafar, R.H., Pappas, G.J., & Hassani, H. (2021). Achieving Linear Convergence in Federated Learning under Objective and Systems Heterogeneity. ArXiv, abs/2102.07053.
[27]H. Zhu, J. Xu, S. Liu, and Y. Jin, “Federated Learning on Non-IID Data: A Survey.” arXiv, Jun. 12, 2021. doi: 10.48550/arXiv.2106.06843.
[28]K. Bonawitz et al., “Towards Federated Learning at Scale: System Design,” Proceedings of Machine Learning and Systems, vol. 1, pp. 374–388, Apr. 2019.
[29]R. Sun et al., “FedMSA: A Model Selection and Adaptation System for Federated Learning,” Sensors, vol. 22, no. 19, p. 7244, Sep. 2022, doi: 10.3390/s22197244.
[30]S. Wang, Y. Li, A. Zhao, and Q. Wang, “Privacy Protection in Federated Learning Based on Differential Privacy and Mutual Information,” 2021 3rd International Conference on Artificial Intelligence and Advanced Manufacture, pp. 428–435, Oct. 2021, doi: 10.1145/3495018.3495093.
[31]X. Gong et al., “Preserving Privacy in Federated Learning with Ensemble Cross-Domain Knowledge Distillation,” AAAI, vol. 36, no. 11, pp. 11891–11899, Jun. 2022, doi: 10.1609/aaai.v36i11.21446.
[32]W. Yang, B. Liu, C. Lu, and N. Yu, “Privacy Preserving on Updated Parameters in Federated Learning,” Proceedings of the ACM Turing Celebration Conference - China, pp. 27–31, May 2020, doi: 10.1145/3393527.3393533.
[33]R. Xu, N. Baracaldo, Y. Zhou, A. Anwar, and H. Ludwig, “HybridAlpha: An Efficient Approach for Privacy-Preserving Federated Learning,” Proceedings of the 12th ACM Workshop on Artificial Intelligence and Security, pp. 13–23, Nov. 2019, doi: 10.1145/3338501.3357371.
[34]T. Nishio and R. Yonetani, “Client Selection for Federated Learning with Heterogeneous Resources in Mobile Edge,” in ICC 2019 - 2019 IEEE International Conference on Communications (ICC), May 2019, pp. 1–7. doi: 10.1109/ICC.2019.8761315.
[35]Kang, Jiawen & Xiong, Zehui & Niyato, Dusit & Yu, Han & Liang, Ying-Chang & Kim, Dong In. (2019). Incentive Design for Efficient Federated Learning in Mobile Networks: A Contract Theory Approach. 1-5. 10.1109/VTS-APWCS.2019.8851649.
[36]L. Li, M. Duan, D. Liu, Y. Zhang, A. Ren, X. Chen, Y. Tan, C. Wang, FedSAE: A Novel Self-Adaptive Federated Learning Framework in Heterogeneous Systems, IJCNN 2021
[37]J. Luo, J. Yang, X. Ye, X. Guo, W. Zhao, FedSkel: Efficient Federated Learning on Heterogeneous Systems with Skeleton Gradients Update, In Proceedings of the 30th ACM International Conference on Information and Knowledge Management (CIKM ’21), pp-(1–5), 2021
[38]B. Luo, W. Xiao, S. Wang, J. Huang, L. Tassiulas, Tackling System and Statistical Heterogeneity for Federated Learning with Adaptive Client Sampling, 2112.11256v1 [cs.LG] , 2021
[39]S. Liu, N. Gupta, and N. H. Vaidya, “Redundancy in cost functions for Byzantine fault-tolerant federated learning,” in Proceedings of the First Workshop on Systems Challenges in Reliable and Secure Federated Learning, Virtual Event Germany: ACM, Oct. 2021, pp. 4–6. doi: 10.1145/3477114.3488761.
[40]S. Bharti and A. Mcgibney, “Privacy-Aware Resource Sharing in Cross-Device Federated Model Training for Collaborative Predictive Maintenance,” IEEE Access, vol. 9, pp. 120367–120379, 2021, doi: 10.1109/ACCESS.2021.3108839.
[41]Y. Jiang et al., “Model Pruning Enables Efficient Federated Learning on Edge Devices.” arXiv, Apr. 06, 2022. Accessed: Jul. 14, 2023. [Online]. Available: http://arxiv.org/abs/1909.12326
[42]Tan, Q., Wang, B., Yu, H., Wu, S., Qian, Y., & Tao, Y. (2023). DP-FEDAW: FEDERATED LEARNING WITH DIFFERENTIAL PRIVACY IN NON-IID DATA. International Journal of Engineering Technologies and Management Research, 10(5), 34–49. https://doi.org/10.29121/ijetmr.v10.i5.2023.1328
[43]T. Tuor, S. Wang, B. J. Ko, C. Liu, and K. K. Leung, “Overcoming Noisy and Irrelevant Data in Federated Learning.” arXiv, Jun. 22, 2020. Accessed: Jul. 14, 2023. [Online]. Available: http://arxiv.org/abs/2001.08300
[44]A. Ghosh, J. Chung, D. Yin, and K. Ramchandran, “An Efficient Framework for Clustered Federated Learning,” in Advances in Neural Information Processing Systems, Curran Associates, Inc., 2020, pp. 19586–19597. Accessed: Jul. 10, 2023. [Online].Available:https://proceedings.neurips.cc/paper_files/paper/2020/hash/e32cc80bf07915058ce90722ee17bb71-Abstract.html4
[45]R. Mishra, H. P. Gupta, and G. Banga, “Resource Aware Clustering for Tackling the Heterogeneity of Participants in Federated Learning.” arXiv, Jun. 07, 2023. Accessed: Jul. 10, 2023. [Online]. Available: http://arxiv.org/abs/2306.04207
[46]D. Zeng, X. Hu, S. Liu, Y. Yu, Q. Wang, and Z. Xu, “Stochastic Clustered Federated Learning.” arXiv, Mar. 01, 2023. Accessed: Jul. 10, 2023. [Online]. Available: http://arxiv.org/abs/2303.00897
[47]Y. Ruan and C. Joe-Wong, “FedSoft: Soft Clustered Federated Learning with Proximal Local Updating.” arXiv, Mar. 22, 2022. Accessed: Jul. 10, 2023. [Online]. Available: http://arxiv.org/abs/2112.06053
[48]C. Briggs, Z. Fan, and P. Andras, “Federated learning with hierarchical clustering of local updates to improve training on non-IID data.” arXiv, May 06, 2020. Accessed: Aug. 20, 2023. [Online]. Available: http://arxiv.org/abs/2004.11791
[49]A. Chen, Y. Fu, L. Wang, and G. Duan, “DWFed: A statistical- heterogeneity-based dynamic weighted model aggregation algorithm for federated learning,” Frontiers in Neurorobotics, vol. 16, 2022, [Online]. Available: https://www.frontiersin.org/articles/10.3389/fnbot.2022.1041553
[50]M. Mendieta, T. Yang, P. Wang, M. Lee, Z. Ding, and C. Chen, “Local Learning Matters: Rethinking Data Heterogeneity in Federated Learning.” arXiv, Apr. 13, 2022. Accessed: Jul. 22, 2023. [Online]. Available: http://arxiv.org/abs/2111.14213
[51]H. B. McMahan, E. Moore, D. Ramage, S. Hampson, and B. A. y Arcas, “Communication-Efficient Learning of Deep Networks from Decentralized Data.” arXiv, Jan. 26, 2023. Accessed: Jul. 22, 2023. [Online]. Available: http://arxiv.org/abs/1602.05629
[52]D. Verma, G. White, S. Julier, S. Pasteris, G. Cirincione, and S. Chakraborty, “Approaches to address the Data Skew Problem in Federated Learning,” May 2019, p. 50. doi: 10.1117/12.2519621.
[53]T. Zhou, Z. Lin, J. Zhang, and D. H. K. Tsang, “Understanding Model Averaging in Federated Learning on Heterogeneous Data.” arXiv, May 20, 2023. doi: 10.48550/arXiv.2305.07845.
[54]“[2206.09979] Mitigating Data Heterogeneity in Federated Learning with Data Augmentation.” https://arxiv.org/abs/2206.09979 (accessed Jul. 23, 2023).
[55]J. Wu, M. Hayat, M. Zhou, and M. Harandi, “Defense against Privacy Leakage in Federated Learning.” arXiv, Sep. 13, 2022. doi: 10.48550/arXiv.2209.05724.
[56]Z. Liu, T. Li, V. Smith, and V. Sekar, “Enhancing the Privacy of Federated Learning with Sketching,” ArXiv, Nov. 2019, Accessed: Jul. 23, 2023. [Online]. Available: https://www.semanticscholar.org/paper/Enhancing-the-Privacy-of-Federated-Learning-with-Liu-Li/cb68e7849eb9ba864926fd869570f04f6ec4edd1#citing-papers
[57]Park, Jaehyoung, and Hyuk Lim. 2022. "Privacy-Preserving Federated Learning Using Homomorphic Encryption" Applied Sciences 12, no. 2: 734. https://doi.org/10.3390/app12020734
[58]Y. Li, Y. Zhou, A. Jolfaei, D. Yu, G. Xu and X. Zheng, "Privacy-Preserving Federated Learning Framework Based on Chained Secure Multiparty Computing," in IEEE Internet of Things Journal, vol. 8, no. 8, pp. 6178-6186, 15 April15, 2021, doi: 10.1109/JIOT.2020.302291
[59]A. Khaled, K. Mishchenko, and P. Richtárik, “Tighter Theory for Local SGD on Identical and Heterogeneous Data.” arXiv, Apr. 14, 2022. Accessed: Aug. 27, 2023. [Online]. Available: http://arxiv.org/abs/1909.04746
[60]T. Li, A. K. Sahu, M. Zaheer, M. Sanjabi, A. Talwalkar, and V. Smith, “Federated Optimization in Heterogeneous Networks.” arXiv, Apr. 21, 2020. Accessed: Aug. 25, 2023. [Online]. Available: http://arxiv.org/abs/1812.06127
[61]S. P. Karimireddy, S. Kale, M. Mohri, S. J. Reddi, S. U. Stich, and A. T. Suresh, “SCAFFOLD: Stochastic Controlled Averaging for Federated Learning.” arXiv, Apr. 09, 2021. Accessed: Sep. 02, 2023. [Online]. Available: http://arxiv.org/abs/1910.06378
[62]Hamer, J., Mohri, M. and Suresh, A.T., 2020, November. Fedboost: A communication-efficient algorithm for federated learning. In International Conference on Machine Learning (pp. 3973-3983). PMLR.
[63]S. Reddi et al., “Adaptive Federated Optimization.” arXiv, Sep. 08, 2021. Accessed: Sep. 02, 2023. [Online]. Available: http://arxiv.org/abs/2003.00295
[64]H. Wang, M. Yurochkin, Y. Sun, D. Papailiopoulos, and Y. Khazaeni, “Federated Learning with Matched Averaging.” arXiv, Feb. 15, 2020. Accessed: Sep. 02, 2023. [Online]. Available: http://arxiv.org/abs/2002.06440
[65]Li, M. Sanjabi, A. Beirami, and V. Smith, “Fair Resource Allocation in Federated Learning.” arXiv, Feb. 14, 2020. Accessed: Aug. 25, 2023. [Online]. Available: http://arxiv.org/abs/1905.10497
[66]J. Wang and G. Joshi, “Adaptive Communication Strategies to Achieve the Best Error-Runtime Trade-off in Local-Update SGD.” arXiv, Mar. 07, 2019. Accessed: Aug. 25, 2023. [Online]. Available: http://arxiv.org/abs/1810.08313
[67]D. Jhunjhunwala, P. Sharma, A. Nagarkatti, and G. Joshi, “FedVARP: Tackling the Variance Due to Partial Client Participation in Federated Learning.” arXiv, Jul. 28, 2022. Accessed: Aug. 27, 2023. [Online]. Available: http://arxiv.org/abs/2207.14130