Submitted by Zihao1 34 Too Good to be Bad: On the Failure of LLMs to Role-Play Villains Tencent 161 5
Submitted by AnnieFeng 25 VeriCoT: Neuro-symbolic Chain-of-Thought Validation via Logical Consistency Checks Amazon Web Services 1
Submitted by taesiri 4 Towards Mitigating Hallucinations in Large Vision-Language Models by Refining Textual Embeddings · 8 authors 1
Submitted by ProKil 3 Real-Time Reasoning Agents in Evolving Environments Social And Language Technology Lab 3 1
Submitted by JiayuJeff 1 CritiCal: Can Critique Help LLM Uncertainty or Confidence Calibration? · 10 authors 1