Distributed optimization method based on L-Katusha and Gossip-mix

21 May 2024, 14:53
12m
107 БК (БиоКорпус, 1 эт.) (МФТИ)

107 БК (БиоКорпус, 1 эт.)

МФТИ

Computer & Data Science 21 Computer & Data Science

Speaker

Ivan Toropin (МФТИ)

Description

Distributed optimization algorithms have emerged as a superior approaches for solving machine learning problems. To accommodate the diverse ways in which data can be stored across devices, these methods must be adaptable to a wide range of situations. As a result, two orthogonal regimes of distributed algorithms are distinguished: horizontal and vertical. During parallel training, communication between nodes can become a critical bottleneck, particularly for high-dimensional and over-parameterized models. Therefore, it is crucial to enhance current methods with strategies that minimize the amount of data transmitted during trainng while still achieving a model of similar quality. This paper introduces a new accelerated algorithm, working in the regime of vertical data division. By utilizing a momentum and variance reduction technique from the Loopless-Katyusha algorithm and Gossip procedure for communications, we provide one of the first theoretical convergence guarantees for the vertical regime.

Primary authors

Aleksandr Beznosikov (Moscow Institute of Physics and Technology) Ivan Toropin (МФТИ) Pyotr Lisov (МФТИ)

Presentation materials