LLM-SRBench is a new benchmark aimed at enhancing scientific equation discovery using large language models, featuring comprehensive evaluation methods and open-source implementation. It includes a structured setup guide for running and contributing new search methods, as well as the necessary configurations for various datasets. The benchmark has been recognized for its significance, being selected for oral presentation at ICML 2025.
LRAGE is an open-source toolkit designed for evaluating Large Language Models in a Retrieval-Augmented Generation context, specifically for legal applications. It integrates various tools and datasets to streamline the evaluation process, allowing researchers to effectively assess model performance with minimal engineering effort. Key features include a modular architecture for retrievers and rerankers, a user-friendly GUI, and support for LLM-as-a-Judge evaluations.