Interpretable Preferences via Multi-Objective Reward Modeling and Mixture-of-Experts

An interpretable reward modeling approach.

May 29, 2024 · 15 min · Haoxiang Wang

Alignment Guidebook

A guidebook for LLM alignment.

March 26, 2024 · 49 min · Shangmin Guo, Wei Xiong