An introductory guide for practitioners interested in finding out how assurance techniques can support the development of responsible AI.
This guide aims to support organisations to better understand how AI assurance techniques can be used to ensure the safe and responsible development and deployment of AI systems. It introduces key AI assurance concepts and terms and situates them within the wider AI governance landscape.
The introduction supports the UK’s March 2023 white paper, A pro-innovation approach to AI regulation that outlines five cross-cutting regulatory principles underpinning AI regulation, and the subsequent consultation response to bring the principles into practice. As AI becomes increasingly prevalent across all sectors of the economy, it is essential that we ensure it is well governed. AI governance refers to a range of mechanisms including laws, regulations, policies, institutions, and norms that can all be used to outline processes for making decisions about AI.
This guidance aims to provide an accessible introduction to both
assurance mechanisms and global technical standards, to help industry
and regulators better understand how to build and deploy responsible AI
systems. The guidance will be regularly updated to reflect feedback from
stakeholders, the changing regulatory environment and emerging global
best practices. (...)
Sem comentários:
Enviar um comentário