Click any tag below to further narrow down your results
Links
Letta agents using a simple filesystem achieve 74.0% accuracy on the LoCoMo benchmark, outperforming more complex memory tools. This highlights that effective memory management relies more on how agents utilize context than on the specific tools employed.
SGI-Bench is a benchmark designed to assess AI systems' capabilities in scientific inquiry, covering stages like deliberation, conception, action, and perception. It includes over 1,000 expert-curated samples from 10 disciplines, focusing on tasks such as deep research, idea generation, and experimental reasoning.
This article benchmarks GPT-5.1, Claude Opus 4.5, and Gemini 3 Pro for security operations tasks. GPT-5.1 and Opus 4.5 show improved accuracy and speed, while Gemini 3 Pro lags behind. The findings help teams choose the best AI model for automation in SecOps.
This article explores the performance of powerful GPUs when paired with a Raspberry Pi compared to traditional desktop PCs. It highlights tests involving media transcoding, 3D rendering, and AI tasks, revealing that the Raspberry Pi can deliver competitive performance at a fraction of the cost and power consumption.
InferenceMAX™ is an open-source automated benchmarking tool that continuously evaluates the performance of popular inference frameworks and models to ensure benchmarks remain relevant amidst rapid software improvements. The platform, supported by major industry players, provides real-time insights into inference performance and is seeking engineers to expand its capabilities.
The Epoch Capabilities Index (ECI) is a composite metric that integrates scores from 39 AI benchmarks into a unified scale for evaluating and comparing model capabilities over time. Utilizing Item Response Theory, the ECI provides a statistical framework to assess model performance against benchmark difficulty, allowing for consistent scoring of AI models such as Claude 3.5 and GPT-5. Future details on the methodology will be published in an upcoming paper funded by Google DeepMind.
The article discusses the fourth day of benchmarking performance for DGX Lab, highlighting the discrepancies between expected results and actual outcomes. It emphasizes the importance of real-world testing in understanding the capabilities of AI hardware and software. The findings aim to inform users about practical applications and performance metrics in AI development.