Search

PKU-Alignment Group @Pair-Lab (under construction)
PKU-Alignment Group @Pair-Lab (under construction)
  • News
  • People
  • Events
  • Publications
  • Contact
  • More Platforms
    知乎 Bilibili Email 小红书 PAIR-Lab
    Copied
    Copied to clipboard
    Kaile Wang

    Kaile Wang

    Ph.D Student

    Ph.D (2026), Peking University

      Interests
      • Reinforcement Learning
      • Safety Alignment
      • LLMs Theory

      Latest

      • AI Alignment: A Comprehensive Survey
      • Safe RLHF-V: Safe Reinforcement Learning from Multi-modal Human Feedback
      • InterMT: Multi-Turn Interleaved Preference Alignment with Human Feedback
      • Reward Generalization in RLHF: A Topological Perspective
      • Stream Aligner: Efficient Sentence-Level Alignment via Distribution Induction
      • Align Anything: Training All-Modality Models to Follow Instructions with Language Feedback
      • Language Models Resist Alignment: Evidence From Data Compression

      © 2025 PKU-Alignment Group.

      Published with Hugo Blox Builder — the free, open source website builder that empowers creators.

      Cite
      Copy Download