ENHANCING PRIVACY AND ACCURACY IN FEDERATED LEARNING FOR REGRESSION WITH SERVER-SIDE FILTERING TO ADDRESS OUTLIERS
DOI:
https://doi.org/10.7494/csci.2026.27.1.7029Abstract
In the contemporary landscape characterized by extensive data proliferation, the amalgamation of information derived from a multitude of devices is imperative for the advanced machine learning models. Nevertheless, the centralization of such data engenders significant apprehensions regarding privacy, particularly when the data is fetched from a heterogeneous array of devices including mobile phones, cameras, sensors, computers, and workstations. Federated Learning proffers a solution to these privacy-related dilemmas by maintaining a decentralized architecture, thereby enabling local devices to preserve their data while concurrently exchanging model parameters. Despite its promise, Federated Learning encounters substantial obstacles concerning data quality, which may arise from inherent biases, the presence of outliers, and the utilization of compromised devices. To mitigate these challenges, we advocate for the implementation of a server-side filtering methodology within Federated Learning, specifically tailored for regression-related problems. Based on this architecture, local devices train the model on their own data sets and then send the learned parameters to a central server. The server is then tasked with the filtration of erroneous contributions, thereby enhancing the overall accuracy of the model. This methodology is substantiated through the application of the Mean Squared Error metric, a widely recognized standard within regression analysis, thereby augmenting both the efficiency and dependability of the learning process while safeguarding user privacy an essential component of Federated Learning.
Downloads
References
[1] Biggio B., Nelson B., Laskov P.: Poisoning attacks against support vector machines. In: arXiv preprint arXiv:1206.6389, 2012.
[2] Blanchard P., El Mhamdi E.M., Guerraoui R., Stainer J.: Machine learning with adversaries: Byzantine tolerant gradient descent. In: Advances in neural information processing systems, vol. 30, 2017.
[3] Bonawitz K., Ivanov V., Kreuter B., Marcedone A., McMahan H.B., Patel S., Ramage D., Segal A., Seth K.: Practical secure aggregation for privacy-preserving machine learning. In: Proceedings of the 2017 ACM SIGSAC Conference on Computer and Communications Security, pp. 1175–1191. 2017.
[4] Falch T.L., Elster A.C.: Machine learning based auto-tuning for enhanced opencl performance portability. In: 2015 IEEE International Parallel and Distributed Processing Symposium Workshop, pp. 1231–1240. IEEE, 2015.
[5] Geyer R.C., Klein T., Nabi M.: Differentially private federated learning: A client level perspective. In: arXiv preprint arXiv:1712.07557, 2017.
[6] Lahitani A.R., Permanasari A.E., Setiawan N.A.: Cosine similarity to determine similarity measure: Study case in online essay assessment. In: 2016 4th International conference on cyber and IT service management, pp. 1–6. IEEE, 2016.
[7] Liu J., Huang J., Zhou Y., Li X., Ji S., Xiong H., Dou D.: From distributed machine learning to federated learning: A survey. In: Knowledge and Information Systems, vol. 64(4), pp. 885–917, 2022.
[8] McMahan B., Moore E., Ramage D., Hampson S., y Arcas B.A.: Communication-efficient learning of deep networks from decentralized data. In: Artificial intelligence and statistics, pp. 1273–1282. PMLR, 2017.
[9] Rivest R.L., Adleman L., Dertouzos M.L., et al.: On data banks and privacy homomorphisms. In: Foundations of secure computation, vol. 4(11), pp. 169-180, 1978.
[10] Steinhardt J., Koh P.W.W., Liang P.S.: Certified defenses for data poisoning attacks. In: Advances in neural information processing systems, vol. 30, 2017.
[11] Tian Y., Zhang W., Simpson A., Liu Y., Jiang Z.L.: Defending against data poisoning attacks: from distributed learning to federated learning. In: The Computer Journal, vol. 66(3), pp. 711–726, 2021.
[12] Wang L., Meng Z., Yang L.: A multi-layer two-dimensional convolutional neural network for sentiment analysis. In: International Journal of Bio-Inspired Computation, vol. 19(2), pp. 97–107, 2022.
[13] Yang Q., Liu Y., Chen T., Tong Y.: Federated machine learning: Concept and applications. In: ACM Transactions on Intelligent Systems and Technology (TIST), vol. 10(2), pp. 1-19, 2019.
[14] Yin D., Chen Y., Kannan R., Bartlett P.: Byzantine-robust distributed learning: Towards optimal statistical rates. In: International conference on machine learning, pp. 5650–5659. Pmlr, 2018.
[15] Zhang X., Fu A., Wang H., Zhou C., Chen Z.: A privacy-preserving and verifiable federated learning scheme. In: ICC 2020-2020 IEEE International Conference on Communications (ICC), pp. 1–6. IEEE, 2020.
Downloads
Published
Issue
Section
License
Copyright (c) 2026 Computer Science

This work is licensed under a Creative Commons Attribution 4.0 International License.