Aligner: Achieving Efficient Alignment through Weak-to-Strong Correction Paper • 2402.02416 • Published Feb 4, 2024 • 4
Reward Generalization in RLHF: A Topological Perspective Paper • 2402.10184 • Published Feb 15, 2024
Kimi k1.5: Scaling Reinforcement Learning with LLMs Paper • 2501.12599 • Published Jan 22 • 127
DeepSeek-V3.2: Pushing the Frontier of Open Large Language Models Paper • 2512.02556 • Published 22 days ago • 228
Language Model Resist Alignment Collection This repository hosts open-sourced models of "Language Model Resist Alignment" (ACL 2025 Main). • 302 items • Updated Jun 11 • 1