Search

Your search keyword '"Richtárik, Peter"' showing total 703 results

Search Constraints

Start Over You searched for: Author "Richtárik, Peter" Remove constraint Author: "Richtárik, Peter"
703 results on '"Richtárik, Peter"'

Search Results

1. Symmetric Pruning of Large Language Models

2. Ringmaster ASGD: The First Asynchronous SGD with Optimal Time Complexity

3. On the Convergence of DP-SGD with Adaptive Clipping

4. MARINA-P: Superior Performance in Non-smooth Federated Optimization with Adaptive Stepsizes

5. Differentially Private Random Block Coordinate Descent

6. Speeding up Stochastic Proximal Optimization in the High Hessian Dissimilarity Setting

7. Methods with Local Steps and Random Reshuffling for Generally Smooth Non-Convex Federated Optimization

8. Pushing the Limits of Large Language Model Quantization via the Linearity Theorem

9. Error Feedback under $(L_0,L_1)$-Smoothness: Normalization and Momentum

10. Tighter Performance Theory of FedExProx

11. Unlocking FedNL: Self-Contained Compute-Optimized Implementation

12. Randomized Asymmetric Chain of LoRA: The First Meaningful Theoretical Framework for Low-Rank Adaptation

13. MindFlayer: Efficient Asynchronous Parallel SGD in the Presence of Heterogeneous and Random Worker Compute Times

14. On the Convergence of FedProx with Extrapolation and Inexact Prox

15. Methods for Convex $(L_0,L_1)$-Smooth Optimization: Clipping, Acceleration, and Adaptivity

16. Cohort Squeeze: Beyond a Single Communication Round per Cohort in Cross-Device Federated Learning

17. Prune at the Clients, Not the Server: Accelerated Sparse Training in Federated Learning

18. SPAM: Stochastic Proximal Point Method with Momentum Variance Reduction for Non-convex Cross-Device Federated Learning

19. A Simple Linear Convergence Analysis of the Point-SAGA Algorithm

20. Local Curvature Descent: Squeezing More Curvature out of Standard and Polyak Gradient Descent

21. On the Optimal Time Complexities in Decentralized Stochastic Asynchronous Optimization

22. A Unified Theory of Stochastic Proximal Point Methods without Smoothness

23. MicroAdam: Accurate Adaptive Optimization with Low Space Overhead and Provable Convergence

24. Freya PAGE: First Optimal Time Complexity for Large-Scale Nonconvex Finite-Sum Optimization with Heterogeneous Asynchronous Computations

25. Stochastic Proximal Point Methods for Monotone Inclusions under Expected Similarity

26. PV-Tuning: Beyond Straight-Through Estimation for Extreme LLM Compression

27. The Power of Extrapolation in Federated Learning

28. FedP3: Federated Personalized and Privacy-friendly Network Pruning under Model Heterogeneity

29. FedComLoc: Communication-Efficient Distributed Training of Sparse and Quantized Models

30. Streamlining in the Riemannian Realm: Efficient Riemannian Optimization with Loopless Variance Reduction

31. LoCoDL: Communication-Efficient Distributed Learning with Local Training and Compression

32. Error Feedback Reloaded: From Quadratic to Arithmetic Mean of Smoothness Constants

33. Improving the Worst-Case Bidirectional Communication Complexity for Nonconvex Distributed Optimization under Function Similarity

34. Shadowheart SGD: Distributed Asynchronous SGD with Optimal Time Complexity Under Arbitrary Computation and Communication Heterogeneity

35. Correlated Quantization for Faster Nonconvex Distributed Optimization

36. Kimad: Adaptive Gradient Compression with Bandwidth Awareness

37. Federated Learning is Better with Non-Homomorphic Encryption

38. Byzantine Robustness and Partial Participation Can Be Achieved at Once: Just Clip Gradient Differences

39. Consensus-Based Optimization with Truncated Noise

40. Communication Compression for Byzantine Robust Learning: New Efficient Algorithms and Improved Rates

41. Variance Reduced Distributed Non-Convex Optimization Using Matrix Stepsizes

42. High-Probability Convergence for Composite and Distributed Stochastic Minimization and Variational Inequalities with Heavy-Tailed Noise

43. Towards a Better Theoretical Understanding of Independent Subnetwork Training

44. Understanding Progressive Training Through the Framework of Randomized Coordinate Descent

45. Improving Accelerated Federated Learning with Compression and Importance Sampling

46. Clip21: Error Feedback for Gradient Clipping

47. Global-QSGD: Practical Floatless Quantization for Distributed Learning with Theoretical Guarantees

48. A Guide Through the Zoo of Biased SGD

49. Error Feedback Shines when Features are Rare

50. Momentum Provably Improves Error Feedback!

Catalog

Books, media, physical & digital resources