← Back to all entries
2026-01-12 🧭 Daily News

Anthropic Open-Sources Evals Framework & Publishes First Annual Transparency Report

Anthropic Open-Sources Evals Framework & Publishes First Annual Transparency Report — visual for 2026-01-12

🧭 Anthropic Open-Sources Its Model Evaluation Framework

Anthropic has released anthropic-evals — the internal evaluation framework used to assess Claude's capabilities, safety properties, and alignment characteristics — as an open-source project under the Apache 2.0 licence. The release includes the evaluation harness, a library of over 400 evaluation tasks covering coding, reasoning, instruction-following, tool use, and safety-relevant scenarios, and the scoring infrastructure used to compute standardised benchmark results for model card publication.

What the framework includes

The framework is available at github.com/anthropics/anthropic-evals. Anthropic encourages contributions and plans to accept community-submitted evaluation tasks that meet quality standards.

evals open source benchmarks AI safety retrospective

🧭 Anthropic Annual Transparency Report 2025 — Safety Activities and Findings

Anthropic has published its first Annual Transparency Report, covering the company's AI safety activities, policy engagements, and model evaluations conducted during 2025. The report is intended as a recurring commitment — Anthropic states it will publish a transparency report every year — and marks the first time the company has consolidated its safety activities across research, policy, and deployment into a single public document.

Highlights from the report

transparency safety annual report Anthropic retrospective