Seminar: Graduate Seminar
Towards Efficient, Robust, and Adaptive Distributed Learning
Distributed learning has become a critical research area due to the rapid growth of data and the demand for scalable Machine Learning (ML). However, ensuring efficient and robust training remains challenging in large-scale systems where data and computation are spread across numerous machines. In this talk, I will present three works that address several of these challenges.
First, I propose an efficient first-order method for stochastic optimization with data originating from a Markov chainโa fundamental problem in peer-to-peer decentralized optimization. This method achieves near-optimal convergence rates while eliminating the need for prior knowledge of the mixing time typically required in previous work. Second, I introduce a downlink compression framework for cross-device Federated Learning, where low-resource clients may participate only intermittently and lack persistent memory, that significantly reduces bandwidth consumption while matching the accuracy of an uncompressed baseline model.ย Finally, I address fault-tolerance in distributed ML (aka Byzantine-robust learning) by developing an algorithm that adapts to dynamic adversaries whose behavior mightย change over timeโunlike previous approaches that assume a fixed set of Byzantine workersโwithout requiring knowledge of the fraction of malicious machines.ย ย
Ph.D. student under the supervision of Prof. Kfir Y. Levy.