PAIR Lab: PKU Alignment and Interaction Research Lab
PAIR Lab: PKU Alignment and Interaction Research Lab
Open-Source Projects
People
Talks
Publications
Resources
Contact
Yifan Zhong
Latest
Panacea: Pareto Alignment via Preference Adaptation for LLMs
Off-agent trust region policy optimization
Heterogeneous-Agent Reinforcement Learning
Maximum entropy heterogeneous-agent reinforcement learning
Safety Gymnasium: A Unified Safe Reinforcement Learning Benchmark
Cite
×