Click any tag below to further narrow down your results
Links
SGI-Bench is a benchmark designed to assess AI systems' capabilities in scientific inquiry, covering stages like deliberation, conception, action, and perception. It includes over 1,000 expert-curated samples from 10 disciplines, focusing on tasks such as deep research, idea generation, and experimental reasoning.
This article benchmarks GPT-5.1, Claude Opus 4.5, and Gemini 3 Pro for security operations tasks. GPT-5.1 and Opus 4.5 show improved accuracy and speed, while Gemini 3 Pro lags behind. The findings help teams choose the best AI model for automation in SecOps.
The Epoch Capabilities Index (ECI) is a composite metric that integrates scores from 39 AI benchmarks into a unified scale for evaluating and comparing model capabilities over time. Utilizing Item Response Theory, the ECI provides a statistical framework to assess model performance against benchmark difficulty, allowing for consistent scoring of AI models such as Claude 3.5 and GPT-5. Future details on the methodology will be published in an upcoming paper funded by Google DeepMind.
ai ✓
benchmarking ✓
models ✓