
arXiv: 2305.02231
handle: 10810/63411 , 10481/84557
Trustworthy Artificial Intelligence (AI) is based on seven technical requirements sustained over three main pillars that should be met throughout the system's entire life cycle: it should be (1) lawful, (2) ethical, and (3) robust, both from a technical and a social perspective. However, attaining truly trustworthy AI concerns a wider vision that comprises the trustworthiness of all processes and actors that are part of the system's life cycle, and considers previous aspects from different lenses. A more holistic vision contemplates four essential axes: the global principles for ethical use and development of AI-based systems, a philosophical take on AI ethics, a risk-based approach to AI regulation, and the mentioned pillars and requirements. The seven requirements (human agency and oversight; robustness and safety; privacy and data governance; transparency; diversity, non-discrimination and fairness; societal and environmental wellbeing; and accountability) are analyzed from a triple perspective: What each requirement for trustworthy AI is, Why it is needed, and How each requirement can be implemented in practice. On the other hand, a practical approach to implement trustworthy AI systems allows defining the concept of responsibility of AI-based systems facing the law, through a given auditing process. Therefore, a responsible AI system is the resulting notion we introduce in this work, and a concept of utmost necessity that can be realized through auditing processes, subject to the challenges posed by the use of regulatory sandboxes. Our multidisciplinary vision of trustworthy AI culminates in a debate on the diverging views published lately about the future of AI. Our reflections in this matter conclude that regulation is a key for reaching a consensus among these views, and that trustworthy and responsible AI systems will be crucial for the present and future of our society.
30 pages, 5 figures, under second review
I.2, FOS: Computer and information sciences, Computer Science - Machine Learning, Computer Science - Artificial Intelligence, responsible AI systems, regulatory sandbox, I.2; K.4; K.5, 68T01, K.4, K.5, Trustworthy AI, Machine Learning (cs.LG), Computer Science - Computers and Society, AI ethics, Artificial Intelligence (cs.AI), trustworthy AI, Computers and Society (cs.CY), AI regulation, Regulatory sandbox, Responsible AI systems
I.2, FOS: Computer and information sciences, Computer Science - Machine Learning, Computer Science - Artificial Intelligence, responsible AI systems, regulatory sandbox, I.2; K.4; K.5, 68T01, K.4, K.5, Trustworthy AI, Machine Learning (cs.LG), Computer Science - Computers and Society, AI ethics, Artificial Intelligence (cs.AI), trustworthy AI, Computers and Society (cs.CY), AI regulation, Regulatory sandbox, Responsible AI systems
| selected citations These citations are derived from selected sources. This is an alternative to the "Influence" indicator, which also reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | 459 | |
| popularity This indicator reflects the "current" impact/attention (the "hype") of an article in the research community at large, based on the underlying citation network. | Top 0.1% | |
| influence This indicator reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | Top 1% | |
| impulse This indicator reflects the initial momentum of an article directly after its publication, based on the underlying citation network. | Top 0.01% |
