view article Article Navigating the RLHF Landscape: From Policy Gradients to PPO, GAE, and DPO for LLM Alignment Feb 11 • 87
Hybrid Reinforcement: When Reward Is Sparse, It's Better to Be Dense Paper • 2510.07242 • Published Oct 8 • 30
Understanding Language Prior of LVLMs by Contrasting Chain-of-Embedding Paper • 2509.23050 • Published Sep 27 • 14
Infusing Theory of Mind into Socially Intelligent LLM Agents Paper • 2509.22887 • Published Sep 26 • 5
LUMINA: Detecting Hallucinations in RAG System with Context-Knowledge Signals Paper • 2509.21875 • Published Sep 26 • 9
Clean First, Align Later: Benchmarking Preference Data Cleaning for Reliable LLM Alignment Paper • 2509.23564 • Published Sep 28 • 7
Understanding Language Prior of LVLMs by Contrasting Chain-of-Embedding Paper • 2509.23050 • Published Sep 27 • 14
Understanding Language Prior of LVLMs by Contrasting Chain-of-Embedding Paper • 2509.23050 • Published Sep 27 • 14 • 2
MetaMind: Modeling Human Social Thoughts with Metacognitive Multi-Agent Systems Paper • 2505.18943 • Published May 25 • 24
DaWin: Training-free Dynamic Weight Interpolation for Robust Adaptation Paper • 2410.03782 • Published Oct 3, 2024 • 1
Towards Calibrated Robust Fine-Tuning of Vision-Language Models Paper • 2311.01723 • Published Nov 3, 2023
Learning Fair Representation via Distributional Contrastive Disentanglement Paper • 2206.08743 • Published Jun 17, 2022
Understanding Multimodal LLMs Under Distribution Shifts: An Information-Theoretic Approach Paper • 2502.00577 • Published Feb 1