Evaluating AI cyber capabilities with crowdsourced elicitation
- ELM
Main:5 Pages
5 Figures
Bibliography:1 Pages
3 Tables
Appendix:2 Pages
Abstract
As AI systems become increasingly capable, understanding their offensive cyber potential is critical for informed governance and responsible deployment. However, it's hard to accurately bound their capabilities, and some prior evaluations dramatically underestimated them. The art of extracting maximum task-specific performance from AIs is called "AI elicitation", and today's safety organizations typically conduct it in-house. In this paper, we explore crowdsourcing elicitation efforts as an alternative to in-house elicitation work.
View on arXivComments on this paper
