Click any tag below to further narrow down your results
Links
The article discusses various open problems in machine learning inspired by a graduate class. It critiques current methodologies, emphasizing the need for a design-based perspective, better evaluation methods, and innovations in large language models. The author encourages researchers to explore these under-addressed areas.
This article discusses how fine-tuning open-source LLM judges using Direct Preference Optimization (DPO) can lead to performance that matches or exceeds GPT-5.2 in evaluating model outputs. The authors trained models like GPT-OSS 120B and Qwen 3 235B on human preference data, achieving better accuracy and efficiency at a lower cost.
Open Deep Research is an open-source agent designed for deep research tasks, compatible with various model providers and search tools. It ranks high on the Deep Research Bench leaderboard and offers flexibility for customization through its API. The platform supports multiple LLMs and search APIs, making it versatile for different research needs.
Bloom is an open source framework that automates the evaluation of AI model behaviors, allowing researchers to specify a desired behavior and generate relevant scenarios for assessment. The tool produces evaluations quickly and offers flexibility in measuring different behavioral traits, complementing existing tools like Petri.
This article details the implementation of Google's Nested Learning (HOPE) architecture, focusing on its mechanism-level components and testing procedures. It provides guidance on installation, usage, and evaluation, including various training configurations and memory management strategies for machine learning models.
LLM-SRBench is a new benchmark aimed at enhancing scientific equation discovery using large language models, featuring comprehensive evaluation methods and open-source implementation. It includes a structured setup guide for running and contributing new search methods, as well as the necessary configurations for various datasets. The benchmark has been recognized for its significance, being selected for oral presentation at ICML 2025.
LRAGE is an open-source toolkit designed for evaluating Large Language Models in a Retrieval-Augmented Generation context, specifically for legal applications. It integrates various tools and datasets to streamline the evaluation process, allowing researchers to effectively assess model performance with minimal engineering effort. Key features include a modular architecture for retrievers and rerankers, a user-friendly GUI, and support for LLM-as-a-Judge evaluations.