Tags: #rlhf

Reinforcement Learning Framework
Ray
9.3k

OpenRLHF/OpenRLHF

An easy-to-use, scalable, and high-performance open-source framework for Reinforcement Learning from Human Feedback (RLHF) based on Ray and vLLM.

LLM Fine-tuning Tool
python
3.7k

hiyouga/ChatGLM-Efficient-Tuning

An efficient toolkit for fine-tuning ChatGLM-6B models using PEFT methods, enabling customization and deployment of large language models.

Educational Resource / Technical Textbook
python
1.8k

natolambert/rlhf-book

A comprehensive open-source textbook and code repository dedicated to Reinforcement Learning from Human Feedback (RLHF) and post-training language models.

LLM Alignment Toolkit
deepspeed
5.6k

huggingface/alignment-handbook

Provides robust recipes and training code to align language models with human and AI preferences, enhancing helpfulness and safety.

Awesome List / Research Resource Collection
4.3k

opendilab/awesome-RLHF

A continually updated, curated list of essential resources for Reinforcement Learning with Human Feedback (RLHF), encompassing research papers, codebases, datasets, and related materials.

Deep Learning Alignment Framework
python
4.6k

PKU-Alignment/align-anything

A modular framework for aligning any-modality large models with human intentions and values using diverse fine-tuning and reinforcement learning methods.

AI/ML Research Framework
Hugging Face
1.6k

PKU-Alignment/safe-rlhf

A modular open-source framework for training constrained value-aligned Large Language Models (LLMs) using Safe Reinforcement Learning from Human Feedback (RLHF).

AI/ML Library
python
1.7k

zai-org/ImageReward

A human preference reward model for evaluating and improving text-to-image generation models.

Machine Learning Research Toolkit
python
1.5k

RLHFlow/RLHF-Reward-Modeling

A comprehensive collection of recipes and code for training various reward models crucial for Reinforcement Learning from Human Feedback (RLHF) in large language models.

LLM Alignment Framework
Python
1.4k

OpenLMLab/MOSS-RLHF

An open-source framework providing code, models, and insights for stable Reinforcement Learning from Human Feedback (RLHF) training in Large Language Models, focusing on the PPO algorithm and reward modeling.

OSS Alternative

Explore the best open source alternatives to commercial software.

© 2026 OSS Alternative. hotgithub.com - All rights reserved.