Ir directamente a la navegación principal Ir directamente a la búsqueda Ir directamente al contenido principal

Connecting the dots in trustworthy Artificial Intelligence: From AI principles, ethics, and key requirements to responsible AI systems and regulation

  • Natalia Díaz-Rodríguez*
  • , Javier Del Ser*
  • , Mark Coeckelbergh
  • , Marcos López de Prado
  • , Enrique Herrera-Viedma
  • , Francisco Herrera
  • *Autor correspondiente de este trabajo
  • University of Granada
  • University of Vienna
  • Cornell University
  • Al Maryah Island
  • Khalifa University of Science and Technology

Producción científica: Contribución a una revistaArtículorevisión exhaustiva

584 Citas (Scopus)
5 Descargas (Pure)

Resumen

Trustworthy Artificial Intelligence (AI) is based on seven technical requirements sustained over three main pillars that should be met throughout the system's entire life cycle: it should be (1) lawful, (2) ethical, and (3) robust, both from a technical and a social perspective. However, attaining truly trustworthy AI concerns a wider vision that comprises the trustworthiness of all processes and actors that are part of the system's life cycle, and considers previous aspects from different lenses. A more holistic vision contemplates four essential axes: the global principles for ethical use and development of AI-based systems, a philosophical take on AI ethics, a risk-based approach to AI regulation, and the mentioned pillars and requirements. The seven requirements (human agency and oversight; robustness and safety; privacy and data governance; transparency; diversity, non-discrimination and fairness; societal and environmental wellbeing; and accountability) are analyzed from a triple perspective: What each requirement for trustworthy AI is, Why it is needed, and How each requirement can be implemented in practice. On the other hand, a practical approach to implement trustworthy AI systems allows defining the concept of responsibility of AI-based systems facing the law, through a given auditing process. Therefore, a responsible AI system is the resulting notion we introduce in this work, and a concept of utmost necessity that can be realized through auditing processes, subject to the challenges posed by the use of regulatory sandboxes. Our multidisciplinary vision of trustworthy AI culminates in a debate on the diverging views published lately about the future of AI. Our reflections in this matter conclude that regulation is a key for reaching a consensus among these views, and that trustworthy and responsible AI systems will be crucial for the present and future of our society.

Idioma originalInglés
Número de artículo101896
PublicaciónInformation Fusion
Volumen99
DOI
EstadoPublicada - nov 2023

Huella

Profundice en los temas de investigación de 'Connecting the dots in trustworthy Artificial Intelligence: From AI principles, ethics, and key requirements to responsible AI systems and regulation'. En conjunto forman una huella única.

Citar esto