dongleecsu 's Collections RLHF papers
updated
Efficient RLHF: Reducing the Memory Usage of PPO
Paper
• 2309.00754
• Published
• 16
Statistical Rejection Sampling Improves Preference Optimization
Paper
• 2309.06657
• Published
• 15
Aligning Large Multimodal Models with Factually Augmented RLHF
Paper
• 2309.14525
• Published
• 32
Stabilizing RLHF through Advantage Model and Selective Rehearsal
Paper
• 2309.10202
• Published
• 11
Aligning Language Models with Offline Reinforcement Learning from Human
Feedback
Paper
• 2308.12050
• Published
• 1
Pairwise Proximal Policy Optimization: Harnessing Relative Feedback for
LLM Alignment
Paper
• 2310.00212
• Published
• 2
A Long Way to Go: Investigating Length Correlations in RLHF
Paper
• 2310.03716
• Published
• 10
Large Language Models Cannot Self-Correct Reasoning Yet
Paper
• 2310.01798
• Published
• 36
Enable Language Models to Implicitly Learn Self-Improvement From Data
Paper
• 2310.00898
• Published
• 24
Paper
• 2309.16609
• Published
• 38
Improving Language Models with Advantage-based Offline Policy Gradients
Paper
• 2305.14718
• Published
• 2
Principle-Driven Self-Alignment of Language Models from Scratch with
Minimal Human Supervision
Paper
• 2305.03047
• Published
• 1