Search

Your search keyword '"Richtarik, Peter"' showing total 335 results

Search Constraints

Start Over You searched for: Author "Richtarik, Peter" Remove constraint Author: "Richtarik, Peter"
335 results on '"Richtarik, Peter"'

Search Results

1. Error Feedback under $(L_0,L_1)$-Smoothness: Normalization and Momentum

2. Tighter Performance Theory of FedExProx

3. Unlocking FedNL: Self-Contained Compute-Optimized Implementation

4. Randomized Asymmetric Chain of LoRA: The First Meaningful Theoretical Framework for Low-Rank Adaptation

5. MindFlayer: Efficient Asynchronous Parallel SGD in the Presence of Heterogeneous and Random Worker Compute Times

6. On the Convergence of FedProx with Extrapolation and Inexact Prox

7. Methods for Convex $(L_0,L_1)$-Smooth Optimization: Clipping, Acceleration, and Adaptivity

8. Cohort Squeeze: Beyond a Single Communication Round per Cohort in Cross-Device Federated Learning

9. Prune at the Clients, Not the Server: Accelerated Sparse Training in Federated Learning

10. SPAM: Stochastic Proximal Point Method with Momentum Variance Reduction for Non-convex Cross-Device Federated Learning

11. A Simple Linear Convergence Analysis of the Point-SAGA Algorithm

12. Local Curvature Descent: Squeezing More Curvature out of Standard and Polyak Gradient Descent

13. On the Optimal Time Complexities in Decentralized Stochastic Asynchronous Optimization

14. A Unified Theory of Stochastic Proximal Point Methods without Smoothness

15. MicroAdam: Accurate Adaptive Optimization with Low Space Overhead and Provable Convergence

16. Freya PAGE: First Optimal Time Complexity for Large-Scale Nonconvex Finite-Sum Optimization with Heterogeneous Asynchronous Computations

17. Stochastic Proximal Point Methods for Monotone Inclusions under Expected Similarity

18. PV-Tuning: Beyond Straight-Through Estimation for Extreme LLM Compression

19. The Power of Extrapolation in Federated Learning

20. FedP3: Federated Personalized and Privacy-friendly Network Pruning under Model Heterogeneity

21. FedComLoc: Communication-Efficient Distributed Training of Sparse and Quantized Models

22. Streamlining in the Riemannian Realm: Efficient Riemannian Optimization with Loopless Variance Reduction

23. LoCoDL: Communication-Efficient Distributed Learning with Local Training and Compression

24. Error Feedback Reloaded: From Quadratic to Arithmetic Mean of Smoothness Constants

25. Improving the Worst-Case Bidirectional Communication Complexity for Nonconvex Distributed Optimization under Function Similarity

26. Shadowheart SGD: Distributed Asynchronous SGD with Optimal Time Complexity Under Arbitrary Computation and Communication Heterogeneity

27. Correlated Quantization for Faster Nonconvex Distributed Optimization

28. Kimad: Adaptive Gradient Compression with Bandwidth Awareness

29. Federated Learning is Better with Non-Homomorphic Encryption

30. Byzantine Robustness and Partial Participation Can Be Achieved at Once: Just Clip Gradient Differences

31. Consensus-Based Optimization with Truncated Noise

32. Communication Compression for Byzantine Robust Learning: New Efficient Algorithms and Improved Rates

33. Variance Reduced Distributed Non-Convex Optimization Using Matrix Stepsizes

34. High-Probability Convergence for Composite and Distributed Stochastic Minimization and Variational Inequalities with Heavy-Tailed Noise

35. Towards a Better Theoretical Understanding of Independent Subnetwork Training

36. Understanding Progressive Training Through the Framework of Randomized Coordinate Descent

37. Improving Accelerated Federated Learning with Compression and Importance Sampling

38. Clip21: Error Feedback for Gradient Clipping

39. Global-QSGD: Practical Floatless Quantization for Distributed Learning with Theoretical Guarantees

40. A Guide Through the Zoo of Biased SGD

41. Error Feedback Shines when Features are Rare

42. Momentum Provably Improves Error Feedback!

43. Explicit Personalization and Local Training: Double Communication Acceleration in Federated Learning

44. Det-CGD: Compressed Gradient Descent with Matrix Stepsizes for Non-Convex Optimization

45. Optimal Time Complexities of Parallel Stochastic Optimization Methods Under a Fixed Computation Model

46. 2Direction: Theoretically Faster Distributed Training with Bidirectional Communication Compression

47. ELF: Federated Langevin Algorithms with Primal, Dual and Bidirectional Compression

48. TAMUNA: Doubly Accelerated Distributed Optimization with Local Training, Compression, and Partial Participation

49. Federated Learning with Regularized Client Participation

50. High-Probability Bounds for Stochastic Optimization and Variational Inequalities: the Case of Unbounded Variance

Catalog

Books, media, physical & digital resources