Abstract
Integrating Artificial Intelligence (AI) and robotics in healthcare is a transformative development with enormous promise for revolutionizing patient care, diagnostics, and treatment modalities. These technologies enhance the precision and efficiency of medical practices, improve patient outcomes, and alleviate the burden on healthcare professionals. While AI offers tremendous potential in healthcare processes, it presents multifaceted ethical challenges that demand meticulous consideration. The primary concerns include bias, privacy, trust, responsibility, transparency, cybersecurity, and data quality.
The appearance of ChatGPT, an artificial intelligence (AI) chatbot with amazing abilities, caused a great sensation in academia. ChatGPT has demonstrated significant potential in various healthcare-related applications, ranging from medical education to clinical decision-making and patient care. Ethical issues brought up using ChatGPT include privacy, authorship, transparency, abuse, and security. One major concern is the potential for bias in ChatGPT’s responses, stemming from biases in the original data given to the chatbot. As AI systems continue to evolve, it is crucial to balance AI advancements and human intervention. While AI has the potential to enhance human intelligence, it is unlikely to completely take over, as human creativity, adaptability, and critical thinking remain indispensable in various aspects of life and scientific research. Establishing clear guidelines and usage policies is essential to ensure the responsible integration of AI in academic and professional settings. Using proactive initiatives and rigorous evaluations can harness AI’s capabilities while upholding stringent ethical standards.
Introduction
The twenty-first century is often recognized as the era of Artificial Intelligence (AI), which raises many questions regarding its impact on human beings. AI is a type of computer system that simulates human intelligence and is used to achieve various tasks by mimicking human cognition, learning, and decision-making processes. It incorporates various technologies and algorithms, including machine learning, deep learning, natural language processing, computer vision, and more.1
Generative AI refers to a subset of AI that generates content, including text and images, by utilizing natural language processing.2 The scope of generative AI reach is both vast and overwhelming, especially when considering the abilities of AI-generated large language models(LLM) such as Chat Generative Pre-trained Transformer (ChatGPT) developed by OpenAI, Google Bard, Microsoft Bing, Claude, and Perplexity.
AI technologies are transforming medicine and healthcare. Scholars and practitioners have debated the ethical, legal, and regulatory implications of medical AI.3The use of AI has enhanced clinical diagnosis, predictive medicine, patients’ data and diagnostics, and clinical decision-making.4
With the launch of ChatGPT, OpenAI has taken the academic community by storm, forcing researchers, editors, and publishers of scientific journals to rethink and adjust their publication policies and strategies. While there are promising results for potential applications of ChatGPT in various fields, there are also significant ethical considerations to be addressed before widespread implementation can prevail.5
Artificial intelligence technologies can also be used in criminal and deceptive activities such as cyber intrusions, electronic fraud, media misinformation, and other illegal and unethical activities. The huge amount of data handled by AI systems could be exposed to hacking or illegal exploitation, which constitutes a serious threat to privacy and security.1
Artificial Intelligence in Healthcare
The emergence of LLM artificial intelligence like ChatGPT, represents one of the most profound developments in healthcare in decades, with the potential to create seismic and revolutionary changes in the practice of medicine.4
The integration of AI and robotics in the healthcare sector has steered into a new era of efficiency and innovation. These technologies offer a multitude of benefits that have the potential to significantly enhance patient care, improve healthcare outcomes, and streamline various healthcare processes.6It enhances diagnosis and treatment capabilities and increases efficiency and productivity in healthcare processes. AI systems provide healthcare professionals with real-time clinical decision support. Robotic-assisted surgeries provide unequaled skill, precision, and stability, reducing the risk of complications and accelerating patient recovery times.AI accelerates drug discovery by analyzing vast datasets to identify potential drug candidates and predict their efficacy.6
The cost savings that AI can bring to the healthcare system is an important driver for incorporating AI technologies. AI applications are estimated to cut annual US healthcare costs by USD 150 billion in 2026. A significant part of these cost reductions stems from changing the healthcare process from a reactive to a proactive approach, focusing on health management rather than disease treatment.7
ChatGPT has demonstrated significant potential in various healthcare-related applications, such as medical education, radiologic decision-making, clinical genetics, patient care, and facilitating communication between patients and healthcare professionals.
Artificial Intelligence in research and medical education
ChatGPT has potential applications in research improving scientific writing, enhancing research versatility, streamlining workflow, saving time, and improving health literacy.8Their utilization comes with potential risks and challenges, including ethical, legal, copyright, transparency, and concerns related to the generation of content difficult to distinguish from human- generated content.9,10The German artist Boris Eldagsen won a photography award but turned it down with the explanation that his image submission was AI-generated and was designed to fool the judges and provoke debate.11 This is a small offence when compared to the way that AI has been used to generate fraudulent images in research publications.12The utilization of ChatGPT poses several other challenges, including bias, plagiarism, lack of originality, inaccurate content, incorrect citations, dehumanization, false forecasting, and the dangers of blind trust.6The reliance on LLMs like ChatGPT for scientific thinking may hinder social and scientific progress, as these models are trained on past data and may not be able to think differently from the past.
A recent systematic review found that ChatGPT improved writing and was useful in healthcare research and education. However, over 95% of responders expressed concerns vis-à-vis ethical and legal issues, copyright, bias and plagiarism, lack of originality, inaccurate content, and security issues.10
ChatGPT is prone to generating fake references and citations, a phenomenon referred to as”hallucination” or “stochastic parroting”.13 This poses a significant challenge for journal editors. The potential misuse of ChatGPT for plagiarism also raises concerns that could disrupt traditional methods of assigning essays and lead to a decline in academic integrity.14
Almost 2% of scientists admitted to data fabrication, falsification, or modification at least once, and over a third admitted to other questionable research practices. Given human nature and fondness for untruth, these are probably conservative estimates.15 There is potentially already a large amount of AI-generated text in medical literature, with more to come unless we are very careful. Indeed, “at a time when trust in science is eroding, it’s important for scientists to recommit to careful and meticulous attention to details.”16
Some have put an outright call for complete rejection of any output produced withAI assistance and others have allowed AI (ChatGPT) to be on the author list.17
Since an AI chatbot is not a human being, in the current legal system, the text automatically generated by an AI chatbot cannot be a copyrighted work; thus, an AI chatbot cannot be an author of an academic paper, not only from the perspective of copyright law but also from the perspective of research ethics. Although researchers can use AI chatbots as research tools, they must be aware that AI chatbots can be competent but dangerous research assistants.18
Ethical Challenges in Artificial Intelligence
Despite its remarkable capabilities, AI has several limitations and ethical implications that need to be considered, particularly in sensitive fields like healthcare and education.13,19,20
Several key issues need to be discussed, including the potential for AI-generated content to be used unethically, the need for transparency and honesty, the risk of manipulating public opinion or decision-making, and the necessity of policies and guidelines.19The literature was primarily focused on the ethics of AI in health care, particularly on diagnostics, and precision medicine, but was largely silent on the ethics of AI in public and population health. The literature highlighted several common ethical concerns related to privacy, bias, trust, accountability, and responsibility. Obviously missing, was the ethics of AI in global health, particularly in the context of low- and middle-income countries.21
The widespread application of AI in the medical field also brings about autonomy and responsibility issues. AI systems may make autonomous decisions in diagnosis and treatment and automatically recommend the best treatment plan without the intervention of a doctor. Such autonomous decision-making may result in the neglect of patients’ interests.1Privacy emerges as a critical concern and ensuring the privacy and security of this data is paramount. Transparency and trust are crucial for successful AI adoption. Cybersecurity concerns are of utmost importance to safeguard patients’ safety and data integrity. Responsibility attribution in AI remains complex and evolving, requiring a balance between human agency and AI capabilities. Shifting from data ownership to data stewardship may ensure responsible data management and privacy protection. The phenomenon of hallucinations in large language models necessitates rigorous validation and fact-checking to ensure the accuracy and reliability of AI-generated outputs.22
Privacy
AI-based applications have a direct impact on patients’ privacy and confidentiality. The loss of control over data access may have a serious psychological impact on patients if their private health information is exposed. The availability of databases involving genetic sequences and medical history could hinder the collection of data and the advancements in medical tests.22
Bias and Inequality
AI systems (especially hu data systems) may contain biases in the way they represent and treat individuals and groups, leading to unfair treatment. There is currently a disparity for black workers. A study in 2019 found that only 2.5% of Google’s employees were black, while Microsoft and Facebook had only 4% representation.23An example of discrimination and racism caused by AI is the “Tay bot” which was launched by Microsoft via Twitter on March 23, 2016; It caused harsh criticism of Microsoft after the bot started posting racist tweets. The developers did not consider the moral risks to the community when the Tay chatbot spread hate and unleashed its tweets, leading Microsoft to shut it down within 24 hours.24
Studies have also revealed poorer implementation rates for specific diseases in rural areas, racial and ethnic minority groups, those without insurance, as well as individuals with lower education and income.22,25
Transparency and Trust
The transparency of the algorithm enables healthcare professionals to understand how ChatGPT formulates its recommendations. Despite existing guidance for transparent reporting, poorly reported medical AI models are still common. Failure to prioritize explainability in clinical decision support systems can jeopardize core ethical values in medicine and may have adverse effects on both individual and public health.26,27
Responsibility and Accountability
AI responsibility attribution poses significant questions regarding who should be held liable for the outcomes of AI actions. The use of AI systems might result in a loss of accountability. Who is responsible for the decisions taken by the AI system, especially when errors are made and harm is done? Are there decisions that AI systems should never make? Should we require algorithmic accountability and transparency? Should we require that the actions of AI systems are always explainable? If an expert medical diagnosis system exists, and kills a patient with an incorrect diagnosis, who is at fault?
Some papers explore human responsibility concerning AI systems. Others advocate for examining the causal chain of human agency, including interactions with technical components like sensors and software, to determine accountability. Shifting from data ownership to data stewardship is crucial to ensure responsible data management, safeguard patients’ privacy, and adhere to regulatory standards. Data stewardship involves governance and protection of data, including determining access and sharing permissions, ensuring regulatory compliance, and facilitating collaborations and data exchange for research and technological advancements.22
Cybersecurity
Cybersecurity is the practice of preventing unauthorized access, theft, damage, or other harmful attacks on computer systems, networks, and digital information. Security breaches can be concealed by AI systems’ incapacity to be explained and interpreted.
Impact on healthcare professionals
The incorporation of AI and robotics into healthcare system not only transforms patient care but also reshapes the roles and responsibilities of healthcare professionals. Automating specific tasks may raise concerns about job displacement among healthcare professionals. Ethical considerations involve ensuring a smooth transition for affected individuals and providing retraining opportunities.6
Conclusion
The use of AI in healthcare introduced unique ethical considerations that demand careful examination and thoughtful resolution. It raises questions about authenticity, accountability, privacy, and security. ChatGPT has shown significant potential in revolutionizing various fields, including science, healthcare, and education, by accelerating processes, enhancing personalization, and providing valuable support to professionals and learners alike.
Despite promising applications, ChatGPT confronts limitations, including critical thinking tasks and generating false references, necessitating stringent cross- verification. For effective and ethical AI deployment, collaboration amongst AI developers, researchers, educators, and policymakers is vital. These tools should augment, not supplant, human expertise. The fusion of technology and healthcare holds vast promise, but only if we navigate its intricacies with conscientiousness and diligence.28 By developing regulatory frameworks and comprehensive guidelines, AI can transform the healthcare process and improve patient outcomes while respecting ethical principles.
AI is likely to be the best or worst thing to ever happen to humanity, so “there is a lot of value in getting it right.” This was the appeal made by Stephen Hawking (1942- 2018) who tried to summarize his call to “humanize artificial intelligence.”28 The international community must “humanize” it, that is, make it serve humans through binding international agreements, otherwise its consequences will be disastrous for humanity.
Unless we are careful and before long, we won’t trust anything for sure and we’ll believe in nothing. What a time to be alive”.29,30,31 With great power and inventions comes great wisdom and responsibility. We bear a significant burden, and our judgment must be spot-on, otherwise, we doom ourselves to disaster.
References
1.Chen C, Chen Z, Luo W, Xu Y, Yang S, Yang G, Chen X, Chi X, Xie N, Zeng Z. Ethical perspective on AI hazards to humans: A review. Medicine (Baltimore). 2023 Dec 1;102(48):e36163.
2.Miao J, Thongprayoon C, Suppadungsuk S, Garcia Valencia OA, Qureshi F, Cheungpasitporn W. Ethical Dilemmas in Using AI for Academic Writing and an Example Framework for Peer Review in Nephrology Academia: A Narrative Review. Clin Pract. 2023 Dec 30;14(1):89-105.
3.Tang L, Li J, Fantus S. Medical artificial intelligence ethics: A systematic review of empirical studies. Digit Health. 2023 Jul 6;9:20552076231186064
4.Shumway DO, Hartman HJ. Medical malpractice liability in large language model artificial intelligence: legal review and policy recommendations. J Osteopath Med. 2024 Jan 31. doi: 10.1515/jom-2023-0229.
5.Watters C, Lemanski MK. Universal skepticism of ChatGPT: a review of early literature on chat generative pre-trained transformer. Front Big Data. 2023 Aug 23;6:1224976.
6.Elendu C, Amaechi DC, Elendu TC, Jingwa KA, Okoye OK, John Okah M, Ladele JA, Farah AH, Alimi HA. Ethical implications of AI and robotics in healthcare: A review. Medicine (Baltimore). 2023 Dec 15;102(50):e36671.
7.Bohr A, Memarzadeh K. The rise of artificial intelligence in healthcare applications. Artificial Intelligence in Healthcare. 2020:25–60. doi: 10.1016/B978-0-12-818438-7.00002-2.
8. Jeyaraman M, Ramasubramanian S, Balaji S, Jeyaraman N, Nallakumarasamy A, Sharma S. ChatGPT in action: Harnessing artificial intelligence potential and addressing ethical challenges in medicine, education, and scientific research. World J Methodol. 2023 Sep 20;13(4):170-178.
9.No authors listed. Tools such as ChatGPT threaten transparent science; here are our ground rules for their use. Nature 2023; 613: 612 [PMID: 36694020
10.Sallam M. ChatGPT Utility in Healthcare Education, Research, and Practice: Systematic Review on the Promising Perspectives and Valid Concerns. Healthcare (Basel). 2023 Mar 19;11(6):887.
11.The Guardian. https://www.theguardian.com/technology/2023/apr/1 7/
12.Gu J, Wang X, Li C, Zhao J, Fu W, Liang G, Qiu J. AI-enabled image fraud in scientific publications. Patterns (N Y). 2022 Jul 8;3(7):100511.
13.Curtis N; ChatGPT. To ChatGPT or not to ChatGPT? The Impact of Artificial Intelligence on Academic Publishing. Pediatr Infect Dis J 2023; 42: 275 [PMID: 36757192
14.King MR; chatGPT. A Conversation on Artificial Intelligence, Chatbots, and Plagiarism in Higher Education. Cell Mol Bioeng 2023; 16: 1-2
15.Grech V, Cuschieri S, Eldawlatly AA. Artificial intelligence in medicine and research – the good, the bad, and the ugly. Saudi J Anaesth. 2023 Jul- Sep;17(3):401-406.
16.Thorp HH. ChatGPT is fun, but not an author. Sience 2023;379:313.
17. Shah PS, Acharya G. Artificial intelligence/machine learning and journalology: Challenges and opportunities. Acta Obstet Gynecol Scand. 2024 Feb;103(2):196-198
18. Lee JY. Can an artificial intelligence chatbot be the author of a scholarly article? J Educ Eval Health Prof. 2023;20:6.
19. van Dis EAM, Bollen J, Zuidema W, van Rooij R, Bockting CL. ChatGPT: five priorities for research. Nature 2023; 614: 224-226.PMID:36737653.
20. Guleria A, Krishan K, Sharma V, Kanchan T. ChatGPT: ethical concerns and challenges in academics and research. J Infect Dev Ctries. 2023 Sep 30;17(9):1292-1299.
21.Murphy K, Di Ruggiero E, Upshur R, Willison DJ, Malhotra N, Cai JC, Malhotra N, Lui V, Gibson J. Artificial intelligence for good health: a scoping review of the ethics literature. BMC Med Ethics. 2021 Feb 15;22(1):14.
22.Jeyaraman M, Balaji S, Jeyaraman N, Yadav S. Unraveling the Ethical Enigma: Artificial Intelligence in Healthcare. Cureus. 2023 Aug 10;15(8):e43262.
23.Jackson MC. Artificial intelligence & algorithmic bias: the issues with technology reflecting history & humans. J Bus Technol Law.2021, 16:299
24.Neff G , Nagy P, Talking to Bots: Symbiotic Agency and the Case of Tay. Intern J Commu,2016,10, 4915– 4931
25.Khoury MJ, Bowen S, Dotson WD, et al.: Health equity in the implementation of genomics and precisionmedicine: a public health imperative. Genet Med. 2022, 24:1630-9.
26.Obasa AE, Palk AC: Responsible application of artificial intelligence in health care . S Afr J Sci. 2023,119:10.17159/sajs.2023/14889
27.Wang C, Liu S, Yang H, Guo J, Wu Y, Liu J. Ethical Considerations of Using ChatGPT in Health Care. J Med Internet Res. 2023 Aug 11;25:e48009.
28.Upadhyay U, Gradisek A, Iqbal U, Dhar E, Li YC, Syed-Abdul S. Call for the responsible artificial intelligence in the healthcare. BMJ Health Care Inform. 2023 Dec 21;30 (1) :e100920.
29.https://www.theguardian.com/science/2016/oct/19/st ephen-hawking-ai-best-or-worst-thing-for-humanity- cambridge
30.Chiles A. In this era of AI photography, I no longer believe my eyes | Adrian Chiles | The Guardian. The Guardian. 2023. Available from: https://www.theguardian.com/media/commentisfree/ 2023/apr/20/
31.Caprioglio A, Paglia L. Fake academic writing: ethics during chatbot era. Eur J Paediatr Dent. 2023 Jun 1;24(2):88-8