Prati
Konstantin Mishchenko
Konstantin Mishchenko
Potvrđena adresa e-pošte na samsung.com - Početna stranica
Naslov
Citirano
Citirano
Godina
Tighter Theory for Local SGD on Identical and Heterogeneous Data
A Khaled, K Mishchenko, P Richtarik
International Conference on Artificial Intelligence and Statistics, 4519-4529, 2020
2682020
Distributed Learning with Compressed Gradient Differences
K Mishchenko, E Gorbunov, M Takáč, P Richtárik
arXiv preprint arXiv:1901.09269, 2019
1412019
Stochastic distributed learning with gradient quantization and double-variance reduction
S Horváth, D Kovalev, K Mishchenko, P Richtárik, S Stich
Optimization Methods and Software, 1-16, 2022
1262022
First Analysis of Local GD on Heterogeneous Data
A Khaled, K Mishchenko, P Richtárik
NeurIPS FL Workshop, arXiv preprint arXiv:1909.04715, 2019
1082019
Random Reshuffling: Simple Analysis with Vast Improvements
K Mishchenko, A Khaled, P Richtárik
Advances in Neural Information Processing Systems 33, 2020
582020
SEGA: Variance Reduction via Gradient Sketching
F Hanzely, K Mishchenko, P Richtárik
Advances in Neural Information Processing Systems, 2082-2093, 2018
552018
Revisiting Stochastic Extragradient
K Mishchenko, D Kovalev, E Shulgin, P Richtárik, Y Malitsky
AISTATS 2020, 2019
482019
Adaptive gradient descent without descent
Y Malitsky, K Mishchenko
International Conference on Machine Learning, 6702-6712, 2020
452020
A Delay-tolerant Proximal-Gradient Algorithm for Distributed Learning
K Mishchenko, F Iutzeler, J Malick, MR Amini
International Conference on Machine Learning, 3584-3592, 2018
352018
Stochastic Newton and cubic Newton methods with simple local linear-quadratic rates
D Kovalev, K Mishchenko, P Richtárik
NeurIPS Workshop Beyond First Order Methods in ML, arXiv preprint arXiv:1912 …, 2019
322019
99% of worker-master communication in distributed optimization is not needed
K Mishchenko, F Hanzely, P Richtárik
Conference on Uncertainty in Artificial Intelligence, 979-988, 2020
27*2020
A distributed flexible delay-tolerant proximal gradient algorithm
K Mishchenko, F Iutzeler, J Malick
SIAM Journal on Optimization 30 (1), 933-959, 2020
242020
Dualize, split, randomize: Toward fast nonsmooth optimization algorithms
A Salim, L Condat, K Mishchenko, P Richtárik
Journal of Optimization Theory and Applications 195 (1), 102-130, 2022
22*2022
ProxSkip: Yes! Local Gradient Steps Provably Lead to Communication Acceleration! Finally!
K Mishchenko, G Malinovsky, S Stich, P Richtárik
International Conference on Machine Learning, 15750-15769, 2022
222022
Proximal and Federated Random Reshuffling
K Mishchenko, A Khaled, P Richtárik
International Conference on Machine Learning, 15718-15749, 2022
172022
A Stochastic Decoupling Method for Minimizing the Sum of Smooth and Non-Smooth Functions
K Mishchenko, P Richtárik
arXiv preprint arXiv:1905.11535, 2019
172019
MISO is Making a Comeback With Better Proofs and Rates
X Qian, A Sailanbayev, K Mishchenko, P Richtárik
arXiv preprint arXiv:1906.01474, 2019
162019
Sinkhorn Algorithm as a Special Case of Stochastic Mirror Descent
K Mishchenko
NeurIPS OTML Workshop, arXiv preprint arXiv:1909.06918, 2019
132019
DAve-QN: A Distributed Averaged Quasi-Newton Method with Local Superlinear Convergence Rate
S Soori, K Mischenko, A Mokhtari, MM Dehnavi, M Gurbuzbalaban
AISTATS 2020, 2019
132019
IntSGD: Adaptive Floatless Compression of Stochastic Gradients
K Mishchenko, B Wang, D Kovalev, P Richtárik
ICLR 2022 - International Conference on Learning Representations, 2022
122022
Sustav trenutno ne može provesti ovu radnju. Pokušajte ponovo kasnije.
Članci 1–20