GitHub / sathishkumar67 / GPT-2-Non-Toxic-RLHF
Aligning GPT2 model to generate Non-Toxic words
JSON API: http://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/sathishkumar67%2FGPT-2-Non-Toxic-RLHF
PURL: pkg:github/sathishkumar67/GPT-2-Non-Toxic-RLHF
Stars: 0
Forks: 0
Open issues: 0
License: mit
Language: Python
Size: 7.93 MB
Dependencies parsed at: Pending
Created at: about 1 year ago
Updated at: about 1 year ago
Pushed at: about 1 year ago
Last synced at: about 1 year ago
Topics: alignment, gpt2, ppo, pytorch, reinforcement-learning, rlhf, text-generation, transformer