Consistency Learning via Decoding Path Augmentation for Transformers in Human Object Interaction Detection Paper • 2204.04836 • Published Apr 11, 2022
Debate or Vote: Which Yields Better Decisions in Multi-Agent Large Language Models? Paper • 2508.17536 • Published Aug 24, 2025
Thinking Makes LLM Agents Introverted: How Mandatory Thinking Can Backfire in User-Engaged Agents Paper • 2602.07796 • Published Feb 8 • 7
When Identity Skews Debate: Anonymization for Bias-Reduced Multi-Agent Reasoning Paper • 2510.07517 • Published 25 days ago
ModeX: Evaluator-Free Best-of-N Selection for Open-Ended Generation Paper • 2601.02535 • Published 25 days ago
Mitigating Selection Bias with Node Pruning and Auxiliary Options Paper • 2409.18857 • Published May 17, 2025
How Contaminated Is Your Benchmark? Quantifying Dataset Leakage in Large Language Models with Kernel Divergence Paper • 2502.00678 • Published May 20, 2025
PICLe: Eliciting Diverse Behaviors from Large Language Models with Persona In-Context Learning Paper • 2405.02501 • Published May 14, 2024
Exploration and Exploitation Errors Are Measurable for Language Model Agents Paper • 2604.13151 • Published 20 days ago • 24
Thinking Makes LLM Agents Introverted: How Mandatory Thinking Can Backfire in User-Engaged Agents Paper • 2602.07796 • Published Feb 8 • 7
Hybrid Reinforcement: When Reward Is Sparse, It's Better to Be Dense Paper • 2510.07242 • Published Oct 8, 2025 • 30