Publication

Distributed Optimization with Byzantine Robustness Guarantees

Lie He
2023
EPFL thesis
Abstract

As modern machine learning continues to achieve unprecedented benchmarks, the resource demands to train these advanced models grow drastically. This has led to a paradigm shift towards distributed training. However, the presence of adversaries—whether malicious or unintentional—complicates the training process. These attacks present notable security and performance challenges. This thesis primarily focuses on enhancing the Byzantine robustness in distributed machine learning. More precisely, we seek to enhance Byzantine robustness across varying conditions, including heterogeneous data, decentralized communication, and preserving input privacy. In this thesis, we formalize these problems and provide solutions backed by theoretical guarantees.Apart from Byzantine robustness, we investigate alternative communication schemes in decentralized learning and methods for improving sample complexities in conditional stochastic optimization (CSO). In decentralized learning, gossip is predominantly the communication technique employed. However, it is susceptible to data heterogeneity and is slow to converge. We introduce a novel relay mechanism implemented over the spanning tree of the communication graph, offering independence of data heterogeneity. Lastly, in addressing the CSO problem, we observe that its stochastic gradient possesses inherent bias stemming from the nested structure of its objective. This bias contributes to an overhead in sample complexity. In this thesis, we enhance the sample complexity by deploying variance reduction and bias correction methods.

About this result
This page is automatically generated and may contain information that is not correct, complete, up-to-date, or relevant to your search query. The same applies to every other page on this website. Please make sure to verify the information with EPFL's official sources.
Related concepts (34)
Machine learning
Machine learning (ML) is an umbrella term for solving problems for which development of algorithms by human programmers would be cost-prohibitive, and instead the problems are solved by helping machines 'discover' their 'own' algorithms, without needing to be explicitly told what to do by any human-developed algorithms. Recently, generative artificial neural networks have been able to surpass results of many previous approaches.
Deep learning
Deep learning is part of a broader family of machine learning methods, which is based on artificial neural networks with representation learning. The adjective "deep" in deep learning refers to the use of multiple layers in the network. Methods used can be either supervised, semi-supervised or unsupervised.
Distributed artificial intelligence
Distributed Artificial Intelligence (DAI) also called Decentralized Artificial Intelligence is a subfield of artificial intelligence research dedicated to the development of distributed solutions for problems. DAI is closely related to and a predecessor of the field of multi-agent systems. Multi-agent systems and distributed problem solving are the two main DAI approaches. There are numerous applications and tools. Distributed Artificial Intelligence (DAI) is an approach to solving complex learning, planning, and decision-making problems.
Show more
Related publications (36)

Optimization Algorithms for Decentralized, Distributed and Collaborative Machine Learning

Anastasiia Koloskova

Distributed learning is the key for enabling training of modern large-scale machine learning models, through parallelising the learning process. Collaborative learning is essential for learning from privacy-sensitive data that is distributed across various ...
EPFL2024

Infusing structured knowledge priors in neural models for sample-efficient symbolic reasoning

Mattia Atzeni

The ability to reason, plan and solve highly abstract problems is a hallmark of human intelligence. Recent advancements in artificial intelligence, propelled by deep neural networks, have revolutionized disciplines like computer vision and natural language ...
EPFL2024

Breaking the Curse of Dimensionality in Deep Neural Networks by Learning Invariant Representations

Leonardo Petrini

Artificial intelligence, particularly the subfield of machine learning, has seen a paradigm shift towards data-driven models that learn from and adapt to data. This has resulted in unprecedented advancements in various domains such as natural language proc ...
EPFL2023
Show more