Reliable and Responsible Foundation Models: A Comprehensive Survey Paper • 2602.08145 • Published 7 days ago • 8
Thinking Makes LLM Agents Introverted: How Mandatory Thinking Can Backfire in User-Engaged Agents Paper • 2602.07796 • Published 3 days ago • 6
Thinking Makes LLM Agents Introverted: How Mandatory Thinking Can Backfire in User-Engaged Agents Paper • 2602.07796 • Published 3 days ago • 6
Thinking Makes LLM Agents Introverted: How Mandatory Thinking Can Backfire in User-Engaged Agents Paper • 2602.07796 • Published 3 days ago • 6
Towards Reducible Uncertainty Modeling for Reliable Large Language Model Agents Paper • 2602.05073 • Published 7 days ago • 11
Towards Reducible Uncertainty Modeling for Reliable Large Language Model Agents Paper • 2602.05073 • Published 7 days ago • 11
Towards Reducible Uncertainty Modeling for Reliable Large Language Model Agents Paper • 2602.05073 • Published 7 days ago • 11
view article Article Navigating the RLHF Landscape: From Policy Gradients to PPO, GAE, and DPO for LLM Alignment Feb 11, 2025 • 106
Hybrid Reinforcement: When Reward Is Sparse, It's Better to Be Dense Paper • 2510.07242 • Published Oct 8, 2025 • 30
Understanding Language Prior of LVLMs by Contrasting Chain-of-Embedding Paper • 2509.23050 • Published Sep 27, 2025 • 15
Infusing Theory of Mind into Socially Intelligent LLM Agents Paper • 2509.22887 • Published Sep 26, 2025 • 6
LUMINA: Detecting Hallucinations in RAG System with Context-Knowledge Signals Paper • 2509.21875 • Published Sep 26, 2025 • 10
Clean First, Align Later: Benchmarking Preference Data Cleaning for Reliable LLM Alignment Paper • 2509.23564 • Published Sep 28, 2025 • 8
Understanding Language Prior of LVLMs by Contrasting Chain-of-Embedding Paper • 2509.23050 • Published Sep 27, 2025 • 15
Understanding Language Prior of LVLMs by Contrasting Chain-of-Embedding Paper • 2509.23050 • Published Sep 27, 2025 • 15 • 2
MetaMind: Modeling Human Social Thoughts with Metacognitive Multi-Agent Systems Paper • 2505.18943 • Published May 25, 2025 • 24