Attestable Audits: Verifiable AI Safety Benchmarks Using Trusted Execution Environments
Schnabl, Christoph, Hugenroth, Daniel, Marino, Bill, Beresford, Alastair R.
–arXiv.org Artificial Intelligence
Benchmarks are important measures to evaluate safety and compliance of AI models at scale. However, they typically do not offer verifiable results and lack confidentiality for model IP and benchmark datasets. We propose Attestable Audits, which run inside Trusted Execution Environments and enable users to verify interaction with a compliant AI model. Our work protects sensitive data even when model provider and auditor do not trust each other. This addresses verification challenges raised in recent AI governance frameworks. We build a prototype demonstrating feasibility on typical audit benchmarks against Llama-3.1.
arXiv.org Artificial Intelligence
Jul-1-2025
- Country:
- Asia > Middle East
- Jordan (0.04)
- Europe > United Kingdom
- England > Cambridgeshire > Cambridge (0.04)
- North America
- Canada (0.04)
- United States
- California > Los Angeles County
- Santa Monica (0.04)
- Colorado (0.04)
- California > Los Angeles County
- Asia > Middle East
- Genre:
- Research Report (0.43)
- Industry:
- Technology: