Click any tag below to further narrow down your results
Links
This article discusses the importance of monitoring the internal reasoning of AI models, rather than just their outputs. It outlines methods for evaluating how effectively this reasoning can be supervised, especially as models become more complex. The authors call for collaborative efforts to enhance the reliability of this monitoring as AI systems scale.
This article presents a codebase for a study on how unified multimodal models (UMMs) enhance reasoning by integrating visual generation. The research introduces a new evaluation suite, VisWorld-Eval, which assesses multimodal reasoning capabilities across various tasks. Experiments show that interleaved visual-verbal reasoning outperforms purely verbal methods in specific contexts.
TextQuests introduces a benchmark to evaluate the performance of Large Language Models (LLMs) in classic text-based video games, focusing on their ability to engage in long-context reasoning and learning through exploration. The evaluation involves assessing agents' progress and ethical behavior across various interactive fiction games, revealing challenges such as hallucination and inefficiency in dynamic thinking. The aim is to help researchers better understand LLM capabilities in complex, exploratory environments.
JudgeLRM introduces a novel approach to using Large Language Models (LLMs) as evaluators, particularly in complex reasoning tasks. By employing reinforcement learning with judge-wise rewards, JudgeLRM models significantly outperform traditional Supervised Fine-Tuning methods and current leading models, demonstrating superior performance in tasks that require deep reasoning.