Search

Your search keyword '"Orvieto, P."' showing total 325 results

Search Constraints

Start Over You searched for: Author "Orvieto, P." Remove constraint Author: "Orvieto, P."
325 results on '"Orvieto, P."'

Search Results

1. Adaptive Methods through the Lens of SDEs: Theoretical Insights on the Role of Noise

2. NIMBA: Towards Robust and Principled Processing of Point Clouds With SSMs

3. Loss Landscape Characterization of Neural Networks without Over-Parametrization

4. Geometric Inductive Biases of Deep Networks: The Role of Data and Architecture

5. An Adaptive Stochastic Gradient Method with Non-negative Gauss-Newton Stepsizes

6. Gradient Descent on Logistic Regression with Non-Separable Data and Large Step Sizes

8. Recurrent neural networks: vanishing and exploding gradients are not the end of the story

9. Understanding the differences in Foundation Models: Attention, State Space Models, and Recurrent Neural Networks

10. On the low-shot transferability of [V]-Mamba

11. Theoretical Foundations of Deep Selective State-Space Models

12. Super Consistency of Neural Network Landscapes and Learning Rate Transfer

13. SDEs for Minimax Optimization

14. Recurrent Distance Filtering for Graph Representation Learning

21. Universality of Linear Recurrences Followed by Non-linear Projections: Finite-Width Guarantees and Benefits of Complex Eigenvalues

22. Achieving a Better Stability-Plasticity Trade-off via Auxiliary Networks in Continual Learning

23. Resurrecting Recurrent Neural Networks for Long Sequences

24. An SDE for Modeling SAM: Theory and Insights

25. An Accelerated Lyapunov Function for Polyak's Heavy-Ball on Convex Quadratics

26. On the Theoretical Properties of Noise Correlation in Stochastic Optimization

27. Mean first exit times of Ornstein-Uhlenbeck processes in high-dimensional spaces

28. Explicit Regularization in Overparametrized Models via Noise Injection

29. Signal Propagation in Transformers: Theoretical Perspectives and the Role of Rank Collapse

31. Dynamics of SGD with Stochastic Polyak Stepsizes: Truly Adaptive Variants and Convergence to Exact Solution

32. Anticorrelated Noise Injection for Improved Generalization

33. On the effectiveness of Randomized Signatures as Reservoir for Learning Rough Dynamics

34. Faster Single-loop Algorithms for Minimax Optimization without Strong Concavity

35. On the Second-order Convergence Properties of Random Search Methods

36. The HERA (Hyper-response Risk Assessment) Delphi consensus for the management of hyper-responders in in vitro fertilization

37. Rethinking the Variational Interpretation of Nesterov's Accelerated Method

38. Vanishing Curvature and the Power of Adaptive Methods in Randomly Initialized Deep Networks

39. Revisiting the Role of Euler Numerical Integration on Acceleration and Stability in Convex Optimization

41. Two-Level K-FAC Preconditioning for Deep Learning

42. Learning explanations that are hard to vary

43. An Accelerated DFO Algorithm for Finite-sum Convex Functions

49. Correction to: The HERA (Hyper‑response Risk Assessment) Delphi consensus for the management of hyper‑responders in in vitro fertilization

50. Momentum Improves Optimization on Riemannian Manifolds

Catalog

Books, media, physical & digital resources