Abstract
Automation of human tasks has taken place for a long time now. Humans have in earlier periods dreamed of a world where machines capable of mimicking decision making would be created with some works of fiction describing in caricature, how machines would take over the human space in the world. Artificial intelligence has come to fruition in the last few decades following the development of fast computing capability and vast chip memory. Discussions of how the human space will look and feel when artificial intelligence have taken place at various levels of global organization geared towards ensuring that the new thinking machines do not rock human society in ways to render them obsolete.
This article looks at the ethics of AI considering the issues that have been outlined by others in the light of communitarian ethics as seen in Africa. It describes the possible impact of thinking machines on society and how individuals would relate with each other and with AI systems.
Author Contributions
Copyright© 2020
K. Langat Simon, et al.
License
This work is licensed under a Creative Commons Attribution 4.0 International License.
This is an open-access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Competing interests The authors have declared that no competing interests exist.
Funding Interests:
Citation:
Introduction
Artificial Intelligence (AI) as we know it, has been around for seven decades. Immediately, following the assembling of electric computers, and probably, owing to old predictions, developers started thinking of new machines that could work more independently of human intervention. Artificial Intelligence became a study discipline attracting many students in universities, pursuing different disciplines from computer science to management, psychology and engineering. This brought about two categories of classifying the studies; theoretical and pragmatic AI accomplishes the decision making without any awareness the way humans would. Thus, it computes rather than think to arrive at a decision. Robots doing different human chores utilize strong or weak AI. Strong AI exhibits general human-like intelligence. Whereas, weak AI mimics human intelligence in dealing with one specific task for which they are developed. Highly specific human characteristics like free will and ethical decision-making is yet to be achieved in AI. Allan Turing predicted in a 1950 proposal that machines would learn until they are indistinguishable from human beings, possibly achieving consciousness. This prediction has not become a reality (depending on what we mean by consciousness) but work continues such that we might see it happen There are several publications on AI around the world but not much on AI and Africa, a developing region with a distinct community and its own experiences and ethic. There are publications on AI and human rights and on AI and libertarian ethics but not any in communitarian ethics (Ubuntu) and AI. This paper singles out ethics of AI with a communitarian approach. It has been noted in a recent publication that Africa has not contributed to the development of regulations that will inform future growth of AI The first computer was switched on in 1946. It dimmed lights in New York City by the colossal (at the time) amount of energy it required. It however had perhaps as much memory as a small calculator would today. Development moved fast and soon industry leaders were thinking about major improvements. In 1955, John McCarthy, Marvin L. Minsky, Nathaniel Rochester and Claude E. Shannon coined the term ‘artificial intelligence’ Our objective is to highlight the ethical/bioethical implications arising from artificial intelligence, showing how it would be perceived and applied in a developing country with a more or less communitarian grounding. At the same time we intend to create awareness on some of the challenging issues amidst the positive effects of artificial intelligence. This paper has a purpose of showing that AI is anthropocentric with man (humans) at the center. We start by giving a brief background, describing briefly the various forms of AI, then the bioethical implications, the legal implications then our recommendations and conclusion. Since antiquity, there have been myths or rumors of humans making artificial ‘beings’ possessing intelligence. These myths were followed by science fiction depictions of intelligent machines performing all sorts of tasks. Some ancient philosophers considered human thinking as mechanical manipulation of symbols. Aristotle (384-322 BC) developed a system of reasoning following simple steps that could lead to decisions. Later, Hobbes would state that inanimate machine objects would be able to follow simple rules and attain reasoning because the process was more or less like computation It is most suitable to define intelligence here, as: “the capacity for logic, understanding, self-awareness, learning, emotional knowledge, reasoning, planning, creativity, critical thinking, and problem solving.” It sounds contradictory to match the two terms artificial and intelligence as the term was coined. Something artificial is a product of human craft. Artificial usually means something insincere, not original, therefore a copy; fake, and inferior to the real. When people talk of artificial rice, eggs, fish it means something negative. AI is interdisciplinary and cross-disciplinary, much like many other areas of engagement today. It involves computer and cognitive science, psychology, philosophy, logic and mathematics (Report of COMEST on robotics ethics no.38). AI in more popular language is “the theory and development of computer systems able to perform tasks normally requiring human intelligence, such as visual perception, speech recognition, decision-making, and translation between languages.” (Wikipedia). As machines become increasingly capable, mental faculties once thought to require intelligence are removed from the definition. Artificial intelligence is not limited to just IT or technology industry, it is found extensively in other areas such as medicine, business, education, law, and manufacturing. https://www.iqvis.com/blog/9-powerful-examples-of-artificial-intelligence-in-use-today/ Humans learn experientially from situations, and AI learns experientially from data. The performance of AI-based machines improves as they receive more data training, much like a person learns through education and experience. The concern with AI is that it appears that humans are surrendering to a paradigm of forced reductionism putting humans into a purely mechanistic, utilitarian model of technology. As AI becomes more and more powerful and invasive, it may inevitably change the world aligning it with the design principles it rests upon. The consequence might be a world full of indistinctive societies. The other worries we can see include; non-benign actors, unconscious and conscious bias informing algorithms and the inevitable enhanced digital divide, manipulation and even coercion, the threat of a new surveillance society with humans turning into super-optimized machines and perhaps the least in a continuum of super-intelligence. AI has a potential to dominate humans or eventually render the species, as we know it, obsolete. One of the trends that came into sharp focus in 2019 was lack of clarity around AI ethics. Harvard University’s Berkman Klein Center sought to extract consensus on AI Ethics, in a report entitled 1. 2. 3. 4. 5. Explainability is the requirement that you be notified that you are interacting with an AI or subjected to an automated decision not involving humans. People should reserve the right to information entitlement letting individuals know about various aspect of, the use of and interaction with AI systems including personal data use in the decision making process. 6. 7. 8. 9. Value based ethics are invariably anthropocentric; humans are central as moral agents and subjects of any acts of man or machine that are morally noteworthy or significant. Communitarian ethics stresses the connection between the individual and the community. The local variant of communitarian ethics is Ubuntu. It focuses more on reciprocity, the common good, tolerance, consensus, mutual respect and value of human life. It defines clearly the relationship between other life forms on earth and man. The advent of AI calls for similar treatment and placement in the earthly collective. Communitarianism asks the questions: what is the social meaning, what are the implications and contexts Hoesle (1992), states that using computerized information systems require people to act and think in prescribed ways that privilege Western cultural traditions because of the origin of computers in these cultures. It contributes in marginalizing the cultural traditions of others Concerns arising from AI include; the widening gap between countries in the economic south and north, colossal job loses in developing countries, increased poverty and loss of rights by minority groups in the same countries. Ubuntu raises distinct issues in the above situation: poverty dehumanises and makes the poor disabled in contributing to the shared life. It lowers the quality of life and increases discontent making some people more of recipients than participants in their communities. This creation of class in new ways will have destabilising effects on society. It is possible that the information the public gets will be manipulated to produce specific social outcomes. This may not be only during election time as has been witnessed in two countries during recent elections; in the US Looking at ethics of AI within an Ubuntu framework provides an opportunity to reexamine the biases that existed prior to AI with renewed vigor. AI can however, contain biases innocently acquired during programming and “learning” that may exhibit some possibly unintended discriminatory effects. Teaching ethics to robots and AI is complicated and has no clear answers much like teaching ethics to children Looking at the eight themes above in the light of Ubuntu as an ethical framework privacy and accountability are key because everyone in a community has to be capable of being a useful member of the community. Each person is therefore accorded respect and autonomy. Safety and security constitute the reason a community is in the first place and therefore Ubuntu shares with other systems like liberalism to importance of these two thematic areas. Since safety and security is expected at all times and are attributed to the source of the act whether human or not, high standards are expected similarly for both the robot or AI system and their human operators or developers. The responsibility for any breach lies with the persons and not the robot or AI system in Ubuntu. Transparency and explainability is important for oversight of operations and beyond that, for regulatory audits. Value based ethics, which include communitarianism places great importance on the character of the subject. To be virtuous in matters of AI means to be transparent and we may add beyond reproach. Developers, implementers and auditors of the AI system must all have to be able to explain in detail the system’s decision-making and predictability. This extends to fairness and non-discrimination in the sense that AI exhibits fairness based on what it has ‘learnt’. It should be noted that it is indeed in health care, banking and criminal justice where there is an enhance sensitivity to and fairness and non-discrimination. Ubuntu here demands extreme caution in this regard. Human control and professional responsibility are two sides of the same coin; leaving AI to its own design is abdicating a responsibility spanning many millennia. Ubuntu would reject this abdication on the account that it does harm to the community.