Search

PKU-Alignment Group @Pair-Lab (under construction)
PKU-Alignment Group @Pair-Lab (under construction)
  • News
  • People
  • Events
  • Publications
  • Contact
  • More Platforms
    知乎 Bilibili Email 小红书 PAIR-Lab
    Copied
    Copied to clipboard
    Juntao Dai

    Juntao Dai

    Research Assistant

    Peking University

      Interests
      • Reinforcement Learning
      • Value Alignment

      Latest

      • AI Alignment: A Comprehensive Survey
      • Safe RLHF-V: Safe Reinforcement Learning from Multi-modal Human Feedback
      • InterMT: Multi-Turn Interleaved Preference Alignment with Human Feedback
      • Sequence to Sequence Reward Modeling: Improving RLHF by Language Feedback
      • OmniSafe: An Infrastructure for Accelerating Safe Reinforcement Learning Research
      • Aligner: Efficient Alignment by Learning to Correct
      • SafeSora: Towards Safety Alignment of Text2Video Generation via a Human Preference Dataset
      • Safety-Gymnasium: A Unified Safe Reinforcement Learning Benchmark
      • BeaverTails: Towards Improved Safety Alignment of LLM via a Human-Preference Dataset

      © 2025 PKU-Alignment Group.

      Published with Hugo Blox Builder — the free, open source website builder that empowers creators.

      Cite
      Copy Download