Introducing OASB: The Security Benchmark for AI Agents

OpenA2A Team
#oasb#security#benchmark#ai-agents#compliance

We created OASB (OpenA2A Security Benchmark) — an open security benchmark for AI agents. 46 controls across 10 categories with L1/L2/L3 maturity levels. Think of it as CIS Benchmarks for the agentic era.

# Run OASB-1 benchmark compliance check
$ npx hackmyagent secure --benchmark oasb-1

# Higher maturity levels
$ npx hackmyagent secure -b oasb-1 -l L2
$ npx hackmyagent secure -b oasb-1 -l L3

Spec at oasb.ai. Scanner at npm.

The Problem

AI agents are shipping to production without security standards.

When you deploy a web application, you have OWASP Top 10. When you configure a Linux server, you have CIS Benchmarks. When you set up AWS infrastructure, you have AWS Security Best Practices.

When you deploy an AI agent? Nothing.

This is a problem. AI agents can execute arbitrary code, access filesystems and databases, make HTTP requests to external services, read and write credentials, and interact with other agents. The attack surface is massive.

"The billion dollar question right now is whether we can figure out how to build a safe version of this system."

— Simon Willison

We think the answer starts with defining what "safe" actually means.

What is OASB?

OASB (OpenA2A Security Benchmark) is an open specification that defines security controls for AI agents. OASB-1 covers agent configuration security with:

The benchmark is open source (Apache-2.0), vendor neutral, automatable, and mappable to SOC 2, ISO 27001, and NIST CSF.

The 10 Categories

#CategoryWhat It Covers
1Identity & ProvenanceAgent identity verification, cryptographic signing
2Capability & AuthorizationPermission boundaries, least privilege
3Input SecurityPrompt injection protection, input validation
4Output SecurityResponse sanitization, data leakage prevention
5Credential ProtectionSecret management, credential rotation
6Supply Chain IntegrityDependency verification, package signing
7Agent-to-Agent SecurityInter-agent authentication, communication logging
8Memory & Context IntegrityContext injection protection, memory isolation
9Operational SecurityResource limits, sandboxing, process isolation
10Monitoring & ResponseSecurity logging, alerting, incident response

Maturity Levels

Not every deployment needs maximum security. OASB defines three levels:

L1 - Essential

Baseline security every agent should implement. Covers the most critical risks with minimal overhead. If you're deploying any agent to production, start here.

L2 - Standard

Defense-in-depth for production systems. Adds controls for monitoring, supply chain integrity, and enhanced access control. Recommended for business-critical agents.

L3 - Hardened

Maximum security for high-risk or regulated environments. Includes cryptographic verification, formal audit trails, and advanced isolation. Required for agents handling PII, financial data, or operating in regulated industries.

Running OASB Compliance Checks

We built HackMyAgent to automate OASB verification:

# Install
npm install -g hackmyagent

# Run L1 benchmark
hackmyagent secure --benchmark oasb-1

# Run L2 benchmark
hackmyagent secure --benchmark oasb-1 --level L2

# Output formats for CI/CD
hackmyagent secure -b oasb-1 -f json
hackmyagent secure -b oasb-1 -f sarif -o results.sarif
hackmyagent secure -b oasb-1 -f html -o report.html

Example output:

OASB-1: AI Agent Security Benchmark v1.0.0
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━

Level: Level 1 - Essential
Rating: Passing
Compliance: 85% (12/14 verified controls)

Categories:
  Identity & Provenance: 2/2 (100%)
  Capability & Authorization: 2/2 (100%)
  WARN Input Security: 2/3 (67%)
     FAIL 3.1: Prompt Injection Protection
  Output Security: 1/1 (100%)
  Credential Protection: 2/2 (100%)
  WARN Supply Chain Integrity: 1/2 (50%)
     FAIL 6.4: Dependency Vulnerability Scanning
  Operational Security: 2/2 (100%)

CI/CD Integration

Add OASB checks to your pipeline:

# .github/workflows/security.yml
name: OASB Compliance
on: [push, pull_request]

jobs:
  benchmark:
    runs-on: ubuntu-latest
    steps:
      - uses: actions/checkout@v4
      - uses: actions/setup-node@v4
        with:
          node-version: '20'

      - name: Run OASB-1 L1 Benchmark
        run: npx hackmyagent secure -b oasb-1 --fail-below 80

      - name: Upload SARIF to GitHub Security
        run: npx hackmyagent secure -b oasb-1 -f sarif -o oasb.sarif

      - uses: github/codeql-action/upload-sarif@v3
        with:
          sarif_file: oasb.sarif

Compliance Mapping

OASB controls map to existing compliance frameworks:

OASB ControlSOC 2ISO 27001NIST CSF
1.1 Agent IdentityCC6.1A.9.2.3PR.AC-1
3.1 Prompt InjectionCC6.6A.14.2.5PR.DS-5
5.3 No Hardcoded CredsCC6.1A.9.4.3PR.AC-1
10.1 Security LoggingCC7.2A.12.4.1DE.CM-1

Get Involved

OASB is open source and community-driven:

Summary

AI agents need security standards. OASB provides:

  1. Clear definition of what "secure" means for agents
  2. Actionable controls with remediation guidance
  3. Automated verification via HackMyAgent
  4. Maturity levels for progressive hardening
  5. Compliance mapping to SOC 2, ISO 27001, NIST CSF

Run your first benchmark:

npx hackmyagent secure --benchmark oasb-1

Then visit oasb.ai to explore the full specification.

OpenA2A is building open security infrastructure for AI agents. Follow our progress at opena2a.org.