Search

Your search keyword '"Yang, Yaodong"' showing total 825 results

Search Constraints

Start Over You searched for: Author "Yang, Yaodong" Remove constraint Author: "Yang, Yaodong"
825 results on '"Yang, Yaodong"'

Search Results

1. Sequence to Sequence Reward Modeling: Improving RLHF by Language Feedback

2. A Survey on Self-play Methods in Reinforcement Learning

3. ProgressGym: Alignment with a Millennium of Moral Progress

4. PKU-SafeRLHF: A Safety Alignment Preference Dataset for Llama Family Models

5. SafeSora: Towards Safety Alignment of Text2Video Generation via a Human Preference Dataset

6. In-Context Editing: Learning Knowledge from Self-Induced Distributions

7. Emerging Safety Attack and Defense in Federated Instruction Tuning of Large Language Models

8. Efficient Adaptation in Mixed-Motive Environments via Hierarchical Opponent Modeling and Planning

9. Language Models Resist Alignment

10. Efficient Model-agnostic Alignment via Bayesian Persuasion

11. Efficient Preference-based Reinforcement Learning via Aligned Experience Estimation

12. Correlated Mean Field Imitation Learning

13. Transforming the Synthesis of Carbon Nanotubes with Machine Learning Models and Automation

14. AnySkill: Learning Open-Vocabulary Physical Skill for Interactive Agents

15. End-to-End Neuro-Symbolic Reinforcement Learning with Textual Explanations

16. Dexterous Functional Pre-Grasp Manipulation with Diffusion Policy

17. Leveraging Team Correlation for Approximating Equilibrium in Two-Team Zero-Sum Games

18. Incentive Compatibility for AI Alignment in Sociotechnical Systems: Positions and Prospects

19. Reward Generalization in RLHF: A Topological Perspective

20. Aligner: Efficient Alignment by Learning to Correct

21. Panacea: Pareto Alignment via Preference Adaptation for LLMs

22. CivRealm: A Learning and Reasoning Odyssey in Civilization for Decision-Making Agents

23. A Perspective of Q-value Estimation on Offline-to-Online Reinforcement Learning

24. JARVIS-1: Open-World Multi-task Agents with Memory-Augmented Multimodal Language Models

25. AI Alignment: A Comprehensive Survey

26. Grasp Multiple Objects with One Hand

27. Safety-Gymnasium: A Unified Safe Reinforcement Learning Benchmark

28. Safe RLHF: Safe Reinforcement Learning from Human Feedback

29. MaskMA: Towards Zero-Shot Multi-Agent Decision Making with Mask-Based Collaborative Learning

30. Robust Multi-Agent Reinforcement Learning by Mutual Information Regularization

31. GEAR: A GPU-Centric Experience Replay System for Large Reinforcement Learning Models

32. ValueDCG: Measuring Comprehensive Human Value Understanding Ability of Language Models

33. Evolving Diverse Red-team Language Models in Multi-round Multi-agent Games

34. Dynamic Handover: Throw and Catch with Bimanual Hands

35. Mixup-Augmented Meta-Learning for Sample-Efficient Fine-Tuning of Protein Simulators

36. ProAgent: Building Proactive Cooperative Agents with Large Language Models

37. JiangJun: Mastering Xiangqi by Tackling Non-Transitivity in Two-Player Zero-Sum Games

38. Theoretically Guaranteed Policy Improvement Distilled from Model-Based Planning

39. SafeDreamer: Safe Reinforcement Learning with World Models

40. BeaverTails: Towards Improved Safety Alignment of LLM via a Human-Preference Dataset

41. Policy Space Diversity for Non-Transitive Games

42. Large Sequence Models for Sequential Decision-Making: A Survey

43. Maximum Entropy Heterogeneous-Agent Reinforcement Learning

44. Deep Reinforcement Learning with Task-Adaptive Retrieval via Hypernetwork

45. Heterogeneous Value Alignment Evaluation for Large Language Models

46. Byzantine Robust Cooperative Multi-Agent Reinforcement Learning as a Bayesian Game

47. OmniSafe: An Infrastructure for Accelerating Safe Reinforcement Learning Research

48. Regret-Minimizing Double Oracle for Extensive-Form Games

49. Heterogeneous-Agent Reinforcement Learning

50. STAS: Spatial-Temporal Return Decomposition for Multi-agent Reinforcement Learning

Catalog

Books, media, physical & digital resources