Search

Your search keyword '"Yang, Zhuoran"' showing total 647 results

Search Constraints

Start Over You searched for: Author "Yang, Zhuoran" Remove constraint Author: "Yang, Zhuoran"
647 results on '"Yang, Zhuoran"'

Search Results

51. Pessimism in the Face of Confounders: Provably Efficient Offline Reinforcement Learning in Partially Observable Markov Decision Processes

52. Embed to Control Partially Observed Systems: Representation Learning with Provable Sample Efficiency

53. Human-in-the-loop: Provably Efficient Preference-based Reinforcement Learning with General Function Approximation

54. Pessimism meets VCG: Learning Dynamic Mechanism Design via Offline Reinforcement Learning

55. Reinforcement Learning from Partial Observation: Linear Function Approximation with Provable Sample Efficiency

59. Learn to Match with No Regret: Reinforcement Learning in Markov Matching Markets

60. The Best of Both Worlds: Reinforcement Learning with Logarithmic Regret and Policy Switches

61. Learning Dynamic Mechanisms in Unknown Environments: A Reinforcement Learning Approach

62. Pessimistic Bootstrapping for Uncertainty-Driven Offline Reinforcement Learning

63. Sequential Information Design: Markov Persuasion Process and Its Efficient Reinforcement Learning

64. Pessimistic Minimax Value Iteration: Provably Efficient Equilibrium Learning from Offline Datasets

65. Joint Differentiable Optimization and Verification for Certified Reinforcement Learning

66. Exponential Family Model-Based Reinforcement Learning via Score Matching

67. Wasserstein Flow Meets Replicator Dynamics: A Mean-Field Analysis of Representation Learning in Actor-Critic

68. Can Reinforcement Learning Find Stackelberg-Nash Equilibria in General-Sum Markov Games with Myopic Followers?

69. ElegantRL-Podracer: Scalable and Elastic Library for Cloud-Native Deep Reinforcement Learning

70. Exponential Bellman Equation and Improved Regret Bounds for Risk-Sensitive Reinforcement Learning

71. False Correlation Reduction for Offline Reinforcement Learning

72. On Reward-Free RL with Kernel and Neural Function Approximations: Single-Agent MDP and Markov Game

73. Optimistic Policy Optimization is Provably Efficient in Non-stationary MDPs

74. Inducing Equilibria via Incentives: Simultaneous Design-and-Play Ensures Global Convergence

78. Provably Efficient Generative Adversarial Imitation Learning for Online and Offline Setting with Linear Function Approximation

79. Online Bootstrap Inference For Policy Evaluation in Reinforcement Learning

80. Towards General Function Approximation in Zero-Sum Markov Games

81. A Unified Off-Policy Evaluation Approach for General Value Function

82. Gap-Dependent Bounds for Two-Player Markov Games

85. Randomized Exploration for Reinforcement Learning with General Value Function Approximation

86. Doubly Robust Off-Policy Actor-Critic: Convergence and Optimality

87. Instrumental Variable Value Iteration for Causal Offline Reinforcement Learning

88. A Near-Optimal Algorithm for Stochastic Bilevel Optimization via Double-Momentum

89. Is Pessimism Provably Efficient for Offline RL?

90. Risk-Sensitive Deep RL: Variance-Constrained Actor-Critic Provably Finds Globally Optimal Policy

91. Variational Transport: A Convergent Particle-BasedAlgorithm for Distributional Optimization

92. On Function Approximation in Reinforcement Learning: Optimism in the Face of Large State Spaces

93. Provable Fictitious Play for General Mean-Field Games

94. Single-Timescale Stochastic Nonconvex-Concave Optimization for Smooth Nonlinear TD Learning

95. Global Convergence of Policy Gradient for Linear-Quadratic Mean-Field Control/Game in Continuous Time

96. Single-Timescale Actor-Critic Provably Finds Globally Optimal Policy

97. Understanding Implicit Regularization in Over-Parameterized Single Index Model

98. A Two-Timescale Framework for Bilevel Optimization: Complexity Analysis and Application to Actor-Critic

99. Provably Efficient Neural Estimation of Structural Equation Model: An Adversarial Approach

100. Dynamic Regret of Policy Optimization in Non-stationary Environments

Catalog

Books, media, physical & digital resources