8 results found Sort:
An Easy-to-use, Scalable and High-performance RLHF Framework (70B+ PPO Full Tuning & Iterative DPO & LoRA & RingAttention & RFT)
Created
2023-07-30
1,219 commits to main branch, last one 17 hours ago
Safe RLHF: Constrained Value Alignment via Safe Reinforcement Learning from Human Feedback
Created
2023-05-15
111 commits to main branch, last one 10 months ago
A simulation framework for RLHF and alternatives. Develop your RLHF method without collecting human data.
Created
2023-05-03
72 commits to main branch, last one about a year ago
Super-Efficient RLHF Training of LLMs with Parameter Reallocation
Created
2024-06-18
1,078 commits to main branch, last one 3 months ago
Okapi: Instruction-tuned Large Language Models in Multiple Languages with Reinforcement Learning from Human Feedback
Created
2023-07-28
35 commits to main branch, last one about a year ago
A repo for RLHF training and BoN over LLMs, with support for reward model ensembles.
Created
2023-12-02
6 commits to main branch, last one 2 months ago
A Survey of Direct Preference Optimization (DPO)
Created
2024-11-26
52 commits to main branch, last one 28 days ago
Official code for ICML 2024 paper, "RIME: Robust Preference-based Reinforcement Learning with Noisy Preferences" (ICML 2024 Spotlight)
Created
2024-04-04
17 commits to main branch, last one 6 months ago