Stochastic Gradient Descent for Spectral Embedding with Implicit Orthogonality Constraint
Graph Chatbot
Chat with Graph Search
Ask any question about EPFL courses, lectures, exercises, research, news, etc. or try the example questions below.
DISCLAIMER: The Graph Chatbot is not programmed to provide explicit or categorical answers to your questions. Rather, it transforms your questions into API requests that are distributed across the various IT services officially administered by EPFL. Its purpose is solely to collect and recommend relevant references to content that you can explore to help you answer your questions.
Data-based control design methods most often consist of iterative adjustment of the controller&psila;s parameters towards the parameter values which minimize an H2 performance criterion. Typically, batches of input-output data collected from the system are ...
A new decomposition optimization algorithm, called path-following gradient-based decomposition, is proposed to solve separable convex optimization problems. Unlike path-following Newton methods considered in the literature, this algorithm does not require ...
Finding convergence rates for numerical optimization algorithms is an important task, because it gives a justification to their use in solving practical problems, while also providing a way to compare their efficiency. This is especially useful in an async ...
Huge scale machine learning problems are nowadays tackled by distributed optimization algorithms, i.e. algorithms that leverage the compute power of many devices for training. The communication overhead is a key bottleneck that hinders perfect scalability. ...
In this paper we elaborate a general homotopy-theoretic framework in which to study problems of descent and completion and of their duals, codescent and cocompletion. Our approach to homotopic (co)descent and to derived (co)completion can be viewed as $\in ...
We propose an algorithmic framework for convex minimization problems of a composite function with two terms: a self-concordant function and a possibly nonsmooth regularization term. Our method is a new proximal Newton algorithm that features a local quadra ...
This work shows how to develop distributed versions of block blind estimation techniques that have been proposed before for batch processing. Using diffusion adaptation techniques, data are accumulated at the nodes to form estimates of the auto-correlation ...
We consider the classical k-means clustering problem in the setting of bi-criteria approximation, in which an algorithm is allowed to output beta*k > k clusters, and must produce a clustering with cost at most alpha times the to the cost of the optimal set ...
Schloss Dagstuhl--Leibniz-Zentrum fuer Informatik2016
Interest in deep probabilistic graphical models has increased in recent years, due to their state-of-the-art perfor- mance on many machine learning applications. Such models are typically trained with the stochastic gradient method, which can take a signif ...
The minimization of empirical risks over finite sample sizes is an important problem in large-scale machine learning. A variety of algorithms has been proposed in the literature to alleviate the computational burden per iteration at the expense of converge ...