Benchmark an AI agent against the production standard.
A PSF-aligned readiness report for agentic systems: self-assessment, optional public GitHub evidence scan, domain scores, critical gaps, and a shareable badge that does not pretend to be certification.
A public report, a transparent methodology, and a README badge give teams something specific to cite while keeping PAI positioned as the standards and lab institution.
Revenue stays downstream
The Index builds trust first. Deploy Studio, Lab reviews, credentials, and partner programs become credible only after the public standard feels real.
Public methodology
Designed for credibility, not vanity scoring.
The first release is deliberately conservative: equal PSF domain weighting, transparent answer values, explicit evidence grading, and a clear line between readiness reports and credentials.
Eight PSF domains
Input boundary, output validation, data stewardship, observability, deployment control, human oversight, security, and ecosystem resilience.
24 checks
Three focused checks per domain, scored as evidence exists, partial, not yet, or not applicable.
Repository signals
Public GitHub scans look for file-path evidence such as evals, schemas, runbooks, approvals, security policy, and fallbacks.
Not certification
The Index is a readiness report. A credential requires separate review, evidence validation, and program governance.