Quantcast
Channel: Machine Learning | Towards AI
Viewing all articles
Browse latest Browse all 792

Pope Francis Talked About AI & Ethics at The G7

$
0
0
Author(s): Harriet Gaywood Originally published on Towards AI. Pope Francis Talked About AI & Ethics at The G7 Credit: Generated by Dall-E 3 This week, Pope Francis addressed the Group of Seven (G7) Summit in Southern Italy about AI and highlighted the importance of ethics as during his address (see full version). He was speaking as part of Italy’s bid for the Presidency of G7. Pope Francis said “At a time when artificial intelligence programs are examining human beings and their actions, it is precisely the ethos concerning the understanding of the value and dignity of the human person that is most at risk in the implementation and development of these systems. Indeed, we must remember that no innovation is neutral. Technology is born for a purpose and, in its impact on human society, always represents a form of order in social relations and an arrangement of power, thus enabling certain people to perform specific actions while preventing others from performing different ones. In a more or less explicit way, this constitutive power dimension of technology always includes the worldview of those who invented and developed it.” He added “We cannot, therefore, conceal the concrete risk, inherent in its fundamental design, that artificial intelligence might limit our worldview to realities expressible in numbers and enclosed in predetermined categories, thereby excluding the contribution of other forms of truth and imposing uniform anthropological, socio-economic and cultural models. The technological paradigm embodied in artificial intelligence runs the risk, then, of becoming a far more dangerous paradigm, which I have already identified as the “technocratic paradigm”. We cannot allow a tool as powerful and indispensable as artificial intelligence to reinforce such a paradigm, but rather, we must make artificial intelligence a bulwark against its expansion.” So what are the other forms of truth? Which parties or stakeholders should be engaged to ensure that AI develops in a way that is ‘beneficial’ to society? This is another way of looking at the question of which organizations are best placed to govern AI and how can we define good decision-making? How can we ensure that AI is ethical or judge what is right and wrong? G7 is united on some issues, but regarding the topic of AI, the focus of its members and other governments aren’t united because their misalignment reflects their individual geopolitical concerns, while global corporations are struggling to take a ‘global approach’ but hitting cultural and national obstacles. So, what kinds of organizations should lead in defining the values that can shape ‘good’ decision-making in AI? The Vatican City’s Rome Call for AI Ethics has attracted signatories, including Microsoft, IBM, Cisco, FAO, and the Pontifical Academy of Life. Under the auspices of the RenAIssance Foundation, established by and run out of the Vatican City, “it aims to promote a sense of responsibility among organizations, governments, institutions and the private sector with the aim to create a future in which digital innovation and technological progress serve human genius and creativity and not their gradual replacement.” This sounds like a laudable and suitably universal ambition that most people would support — but to achieve this, the representation needs to be broad. If one religious organization is leading discussions on this topic, then other religious voices must be included to create a diversity of views. This has happened to some extent — in January 2023, jewish and muslim religious leaders signed the Rome Call. So, which other religions should be included if religious leaders are going to steer the direction and contribute truths? If Italy is successful in its bid for the Presidency of G7, there has already been a hint of what to expect. On September 2023, Italy’s Prime Minister Giorgia Meloni spoke spoke at the UN saying “We cannot make the mistake of considering this domain [AI] a ‘free zone’ without rules. We need global governance mechanisms capable of ensuring that these technologies respect ethical barriers and that the evolution of technology remains at the service of humans and not vice versa. We need to give practical application to the concept of “algorethics,” that is, giving ethics to algorithms”. So, can we define values and ethical behavior that transcend religions? Or will certain governments and politics always dominate? Business leaders are also being challenged on a national level. In April 2024, the US Department of Homeland Security (DHS) created an AI Safety and Security Board, which includes members of corporations, academics, policymakers and civil organizations (see full list of DHS current board members). Corporations include Adobe, Alphabet, Amazon Web Services, AMD, Cisco, Microsoft, and Nvidia. So, how should a global corporation balance its business interests with protecting national security? The US is clear about its concerns. It has released ‘Guidelines to mitigate AI critical infrastructure’ and released a report (link) about AI misuse in the development and production of chemical, biological, radiological, and nuclear (CBRN). This builds on the AI Risk Management Framework (RMF) by the National Institute of Standards and Technology (NIST). If we contrast this with the UK’s AI Bill (just 7 pages), which includes the establishment of an AI Authority and defines AI as follows: (i) In this Act “artificial intelligence” and “AI” mean technology enabling the programming or training of a device or software to — (a) perceive environments through the use of data; (b) interpret data using automated processing designed to approximate cognitive abilities; and © make recommendations, predictions or decisions; with a view to achieving a specific objective. (ii) generative AI, meaning deep or large language models able to generate text and other content based on the data on which they were trained.” This feels very vague and limited in its reach given the number or types of AI not included. A research briefing about the AI Bill in the House of Lords — 18 March, 2024, states “at present, it is too soon in the evolution of AI technology to legislate effectively and to do so now may be counterproductive.” Therefore a ‘voluntary’ approach of self-governance […]

Viewing all articles
Browse latest Browse all 792

Trending Articles