The Alignment Waltz: Jointly Training Agents to Collaborate for Safety Paper • 2510.08240 • Published Oct 9 • 41
IA2: Alignment with ICL Activations Improves Supervised Fine-Tuning Paper • 2509.22621 • Published Sep 26 • 8
view article Article Navigating the RLHF Landscape: From Policy Gradients to PPO, GAE, and DPO for LLM Alignment Feb 11 • 89
Jointly Reinforcing Diversity and Quality in Language Model Generations Paper • 2509.02534 • Published Sep 2 • 24
Feedback Friction: LLMs Struggle to Fully Incorporate External Feedback Paper • 2506.11930 • Published Jun 13 • 53
To CoT or not to CoT? Chain-of-thought helps mainly on math and symbolic reasoning Paper • 2409.12183 • Published Sep 18, 2024 • 39
AnaloBench: Benchmarking the Identification of Abstract and Long-context Analogies Paper • 2402.12370 • Published Feb 19, 2024 • 2
Benchmarking Language Model Creativity: A Case Study on Code Generation Paper • 2407.09007 • Published Jul 12, 2024 • 4
Cognition Collection Perception and abstraction. Each modality is tokenized and embedded into vectors for model to comprehend. • 200 items • Updated Apr 15 • 6
RATIONALYST: Pre-training Process-Supervision for Improving Reasoning Paper • 2410.01044 • Published Oct 1, 2024 • 35