Federated Learning: A Comprehensive Study on Decentralized Machine Learning
DOI:
https://doi.org/10.61808/jsrt183Keywords:
Federated Learning, Non-IID Data, Communication Costs, Data PrivacyAbstract
Federated Learning (FL) represents a paradigm shift in machine learning, enabling collaborative training across decentralized devices while preserving data privacy. This study comprehensively examines FL’s architecture, algorithms, and challenges using the MNIST dataset as a testbed. We compare two prominent algorithms—Federated Averaging (FedAvg) and Federated Proximal (FedProx)—under independent and identically distributed (IID) and non-IID conditions. Results show FedAvg achieving 97.5% accuracy on IID data, while FedProx outperforms it with 94.4% on non-IID data, a 2.1% improvement over FedAvg’s 92.3%. We also assess scalability, communication costs, and privacy vulnerabilities, supported by graphs, tables, and images. This work underscores FL’s potential in privacysensitive applications like healthcare and IoT, identifying key limitations and future research directions.