10 links
tagged with all of: machine-learning + reinforcement-learning
Click any tag below to further narrow down your results
Links
The repository serves as a comprehensive resource for the survey paper "The Landscape of Agentic Reinforcement Learning for LLMs: A Survey," detailing various reinforcement learning methods and their applications to large language models (LLMs). It includes tables summarizing methodologies, objectives, and key mechanisms, alongside links to relevant papers and resources in the field of AI.
CrystalFormer is a transformer-based autoregressive model tailored for generating crystalline materials while adhering to space group symmetry, enhancing data and computational efficiency. It allows for conditional generation through a structured framework, which includes reinforcement learning and Markov chain Monte Carlo methods. The model supports various functionalities such as generating specific crystal structures and evaluating their validity and novelty.
Large language models (LLMs) typically cannot adapt their weights dynamically to new tasks or knowledge. The Self-Adapting LLMs (SEAL) framework addresses this limitation by allowing models to generate their own finetuning data and directives for self-adaptation through a reinforcement learning approach, resulting in persistent weight updates and improved performance in knowledge incorporation and few-shot generalization tasks.
TreeRL is a novel reinforcement learning framework that integrates on-policy tree search to enhance the training of language models. By incorporating intermediate supervision and optimizing search efficiency, TreeRL addresses issues common in traditional reinforcement learning methods, such as distribution mismatch and reward hacking. Experimental results show that TreeRL outperforms existing methods in math and code reasoning tasks, showcasing the effectiveness of tree search in this domain.
Fulcrum Research is developing tools to enhance human oversight in a future where AI agents perform tasks such as software development and research. Their goal is to create infrastructure for safely deploying these agents, focusing on improving machine learning evaluations and environments. They invite collaboration from those working on reinforcement learning and agent deployment.
The article discusses the process of reinforcement learning fine-tuning, detailing how to enhance model performance through specific training techniques. It emphasizes the importance of tailored approaches to improve the adaptability and efficiency of models in various applications. The information is aimed at practitioners looking to leverage reinforcement learning for real-world tasks.
WavReward is a novel reward feedback model designed to evaluate spoken dialogue systems by assessing both their intelligence quotient (IQ) and emotional quotient (EQ) through audio language models. It introduces a specialized evaluator using multi-sample feedback and reinforcement learning, along with the ChatReward-30K dataset, significantly outperforming existing evaluation models in accuracy and subjective testing across various spoken dialogue scenarios.
The article discusses the concept of spurious rewards in reinforcement learning systems, emphasizing the need to rethink training signals for more effective learning outcomes. It highlights the potential pitfalls of relying on misleading rewards that can skew the training process and suggests strategies for improving reward design.
JudgeLRM introduces a novel approach to using Large Language Models (LLMs) as evaluators, particularly in complex reasoning tasks. By employing reinforcement learning with judge-wise rewards, JudgeLRM models significantly outperform traditional Supervised Fine-Tuning methods and current leading models, demonstrating superior performance in tasks that require deep reasoning.
Thyme introduces a groundbreaking approach to image processing by autonomously generating and executing code for complex visual reasoning tasks. Utilizing a two-stage training strategy that combines supervised fine-tuning and reinforcement learning, along with the innovative GRPO-ATS algorithm, it effectively enhances performance in high-resolution perception.