Search

Your search keyword '"Gu, Quanquan"' showing total 563 results

Search Constraints

Start Over You searched for: Author "Gu, Quanquan" Remove constraint Author: "Gu, Quanquan"
563 results on '"Gu, Quanquan"'

Search Results

1. Self-Play Preference Optimization for Language Model Alignment

2. Matching the Statistical Query Lower Bound for k-sparse Parity Problems with Stochastic Gradient Descent

3. Guided Discrete Diffusion for Electronic Health Record Generation

4. Nearly Optimal Algorithms for Contextual Dueling Bandits from Adversarial Feedback

5. Settling Constant Regrets in Linear Markov Decision Processes

6. Feel-Good Thompson Sampling for Contextual Dueling Bandits

7. Antigen-Specific Antibody Design via Direct Energy-based Preference Optimization

8. Protein Conformation Generation via Force-Guided SE(3) Diffusion Models

9. DecompOpt: Controllable and Decomposed Diffusion Models for Structure-based Molecular Optimization

10. Autonomous closed-loop mechanistic investigation of molecular electrochemistry via automation.

11. Causal Graph ODE: Continuous Treatment Effect Modeling in Multi-agent Dynamical Systems

12. Diffusion Language Models Are Versatile Protein Learners

13. DecompDiff: Diffusion Models with Decomposed Priors for Structure-Based Drug Design

14. Self-Play Fine-Tuning of Diffusion Models for Text-to-Image Generation

15. Reinforcement Learning from Human Feedback with Active Queries

16. Nearly Minimax Optimal Regret for Learning Linear Mixture Stochastic Shortest Path

17. Towards Robust Model-Based Reinforcement Learning Against Adversarial Corruption

18. Mitigating Object Hallucination in Large Vision-Language Models via Classifier-Free Guidance

19. TrustLLM: Trustworthiness in Large Language Models

20. Self-Play Fine-Tuning Converts Weak Language Models to Strong Language Models

21. Sparse PCA with Oracle Property

22. Fast Sampling via De-randomization for Discrete Diffusion Models

23. Risk Bounds of Accelerated SGD for Overparameterized Linear Regression

24. Rephrase and Respond: Let Large Language Models Ask Better Questions for Themselves

25. Implicit Bias of Gradient Descent for Two-layer ReLU and Leaky ReLU Networks on Nearly-orthogonal Data

26. Corruption-Robust Offline Reinforcement Learning with General Function Approximation

27. Pure Exploration in Asynchronous Federated Bandits

28. How Many Pretraining Tasks Are Needed for In-Context Learning of Linear Regression?

29. Why Does Sharpness-Aware Minimization Generalize Better Than SGD?

30. Pessimistic Nonlinear Least-Squares Value Iteration for Offline Reinforcement Learning

31. Variance-Aware Regret Bounds for Stochastic Contextual Dueling Bandits

32. Understanding Transferable Representation Learning and Zero-shot Transfer in CLIP

33. Diffusion Language Models Can Perform Many Tasks with Scaling and Instruction-Finetuning

34. The Implicit Bias of Batch Normalization in Linear Models and Two-layer Linear Convolutional Neural Networks

35. Robust Learning with Progressive Data Expansion Against Spurious Correlation

37. Domain Specialization as the Key to Make Large Language Models Disruptive: A Comprehensive Survey

38. Horizon-free Reinforcement Learning in Adversarial Linear Mixture MDPs

39. Uniform-PAC Guarantees for Model-Based RL with Bounded Eluder Dimension

40. Cooperative Multi-Agent Reinforcement Learning: Asynchronous Communication and Linear Function Approximation

41. Personalized Federated Learning under Mixture of Distributions

42. Optimal Horizon-Free Reward-Free Exploration for Linear Mixture MDPs

43. On the Interplay Between Misspecification and Sub-optimality Gap in Linear Contextual Bandits

44. The Benefits of Mixup for Feature Learning

45. Borda Regret Minimization for Generalized Linear Dueling Bandits

46. Benign Overfitting for Two-layer ReLU Convolutional Neural Networks

47. Finite-Sample Analysis of Learning High-Dimensional Single ReLU Neuron

48. Variance-Dependent Regret Bounds for Linear Bandits and Reinforcement Learning: Adaptivity and Computational Efficiency

49. Structure-informed Language Models Are Protein Designers

50. Nearly Minimax Optimal Reinforcement Learning for Linear Markov Decision Processes

Catalog

Books, media, physical & digital resources