3 links
tagged with all of: reasoning + large-language-models
Click any tag below to further narrow down your results
Links
Continued scaling of large language models (LLMs) may not yield diminishing returns as previously thought; even small improvements in accuracy can lead to significant advancements in long-horizon task execution. The study reveals that LLMs struggle with longer tasks not due to reasoning limitations, but execution errors that compound over time, highlighting the importance of model size and strategic thinking in improving performance.
TextQuests introduces a benchmark to evaluate the performance of Large Language Models (LLMs) in classic text-based video games, focusing on their ability to engage in long-context reasoning and learning through exploration. The evaluation involves assessing agents' progress and ethical behavior across various interactive fiction games, revealing challenges such as hallucination and inefficiency in dynamic thinking. The aim is to help researchers better understand LLM capabilities in complex, exploratory environments.
JudgeLRM introduces a novel approach to using Large Language Models (LLMs) as evaluators, particularly in complex reasoning tasks. By employing reinforcement learning with judge-wise rewards, JudgeLRM models significantly outperform traditional Supervised Fine-Tuning methods and current leading models, demonstrating superior performance in tasks that require deep reasoning.