Skip to content

PersonaTeaming: Exploring How Introducing Personas Can Improve Automated AI Red-Teaming Apple Machine Learning Research

​[[{“value”:”This paper was accepted at the Workshop on Regulatable ML (ReML) at NeurIPS 2025.
Recent developments in AI governance and safety research have called for red-teaming methods that can effectively surface potential risks posed by AI models. Many of these calls have emphasized how the identities and backgrounds of red-teamers can shape their red-teaming strategies, and thus the kinds of risks they are likely to uncover. While automated red-teaming approaches promise to complement human red-teaming by enabling larger-scale exploration of model behavior, current approaches do not consider the role…”}]] [[{“value”:”This paper was accepted at the Workshop on Regulatable ML (ReML) at NeurIPS 2025.
Recent developments in AI governance and safety research have called for red-teaming methods that can effectively surface potential risks posed by AI models. Many of these calls have emphasized how the identities and backgrounds of red-teamers can shape their red-teaming strategies, and thus the kinds of risks they are likely to uncover. While automated red-teaming approaches promise to complement human red-teaming by enabling larger-scale exploration of model behavior, current approaches do not consider the role…”}]]  Read More  

Leave a Reply

Your email address will not be published. Required fields are marked *