194

Evaluating AI cyber capabilities with crowdsourced elicitation

Main:5 Pages
5 Figures
Bibliography:1 Pages
3 Tables
Appendix:2 Pages
Abstract

As AI systems become increasingly capable, understanding their offensive cyber potential is critical for informed governance and responsible deployment. However, it's hard to accurately bound their capabilities, and some prior evaluations dramatically underestimated them. The art of extracting maximum task-specific performance from AIs is called "AI elicitation", and today's safety organizations typically conduct it in-house. In this paper, we explore crowdsourcing elicitation efforts as an alternative to in-house elicitation work.

View on arXiv
Comments on this paper