Change the repository type filter
All
Repositories list
23 repositories
- Align Anything: Training All-modality Model with Feedback
- Safe RLHF: Constrained Value Alignment via Safe Reinforcement Learning from Human Feedback
- NeurIPS 2023: Safety-Gymnasium: A Unified Safe Reinforcement Learning Benchmark
eval-anything
Publicllms-resist-alignment
PublicSafeVLA
Publicomnisafe
PublicJMLR: OmniSafe is an infrastructural framework for accelerating SafeRL research.ProAgent
PublicAAAI24(Oral) ProAgent: Building Proactive Cooperative Agents with Large Language ModelsBeaver-zh-hk
PublicTransformerLens-V
PublicSAELens-V
Publicaligner
Public.github
PublicAligner2024.github.io
Publicsafe-sora
PublicSafeSora is a human preference dataset designed to support safety alignment research in the text-to-video generation field, aiming to enhance the helpfulness and harmlessness of Large Vision Models (LVMs).SafeDreamer
PublicICLR 2024: SafeDreamer: Safe Reinforcement Learning with World ModelsSafe-Policy-Optimization
PublicNeurIPS 2023: Safe Policy Optimization: A benchmark repository for safe reinforcement learning algorithmsAlignmentSurvey
PublicAI Alignment: A Comprehensive Surveybeavertails
PublicReDMan
Public