Ecosyste.ms: Repos
An open API service providing repository metadata for many open source software ecosystems.
GitHub topics: safe-rlhf
PKU-Alignment/safe-rlhf
Safe RLHF: Constrained Value Alignment via Safe Reinforcement Learning from Human Feedback
Language: Python - Size: 4.01 MB - Last synced: about 2 months ago - Pushed: about 2 months ago - Stars: 1,137 - Forks: 92
EzgiKorkmaz/adversarial-reinforcement-learning
Reading list for adversarial perspective and robustness in deep reinforcement learning.
Size: 15.6 KB - Last synced: about 2 months ago - Pushed: 9 months ago - Stars: 74 - Forks: 3
PKU-Alignment/beavertails
BeaverTails is a collection of datasets designed to facilitate research on safety alignment in large language models (LLMs).
Language: Makefile - Size: 2.33 MB - Last synced: 7 months ago - Pushed: 7 months ago - Stars: 42 - Forks: 1