metricas
covid
Buscar en
Journal of Innovation & Knowledge
Toda la web
Inicio Journal of Innovation & Knowledge Generative AI for decision-making: A multidisciplinary perspective
Journal Information
Visits
640
Vol. 10. Issue 4. (In progress)
(July - August 2025)
Full text access
Generative AI for decision-making: A multidisciplinary perspective
Visits
640
Mousa Albashrawia,b,
Corresponding author
bishrama@kfupm.edu.sa

Correspondence to. IRC for Finance and Digital Economy, KFUPM, Saudi Arabia.
a IRC for Finance and Digital Economy, KFUPM, Saudi Arabia
b Information Systems & Operations Management Department, KFUPM, Saudi Arabia
This item has received
Article information
Abstract
Full Text
Bibliography
Download PDF
Statistics
Figures (2)
Tables (2)
Table 1. Generative AI research on decision-making: themes, descriptions, and citations.
Tables
Table 2. Comparative analysis of key themes in GenAI-driven decision-making across domains.
Tables
Show moreShow less
Abstract

Generative artificial intelligence (GenAI) is rapidly reshaping decision-making across multiple domains, including health, law, business, education, and tourism. This study synthesizes the fragmented research on GenAI to provide a comprehensive framework for understanding its role in enhancing decision-making accuracy, efficiency, and personalization. Employing a systematic literature review and thematic analysis, this study categorizes diverse applications, from clinical diagnostics and legal reasoning to financial advisement and educational support, highlighting both innovative practices and persistent challenges. The analysis of 101 articles reveals that, while GenAI significantly improves data processing and decision support, mitigating issues such as inherent bias, misinformation, and transparency deficits requires careful attention. The integration of multi-agent frameworks and human oversight is critical for ensuring ethical and reliable outcomes. Ultimately, this synthesis highlights the transformative potential of GenAI as a decision-making tool by presenting a cross-disciplinary framework that reveals its impact and uncovers gaps across various domains. The study also advocates the development of robust regulatory and technological strategies to harness the benefits and address the limitations of GenAI.

Keywords:
GenAI
Decisions making
Health
Responsible AI
Ethical governance
JEL classification:
M0
M1
O3
O4
Abbreviations:
AHP
GenAI
LLM
RRR
SKIG
Full Text
Introduction

Generative artificial intelligence (GenAI) has emerged as a groundbreaking innovation with the potential to reshape decision-making across multiple domains. Advances in machine learning and natural language processing have enabled models such as ChatGPT to assist in various decision-making processes, ranging from clinical diagnostics and legal adjudication to financial advisement and educational planning (Abdelwahed, 2024; Cardoso et al., 2024). The rapid integration of GenAI into these fields highlights its significance as both a practical tool for enhancing efficiency and an influential force driving transformative change. According to McKinsey, 75 % of professionals expect GenAI to cause significant or disruptive change in their industry within the next three years, underscoring its growing role in strategic decision-making and business innovation (Marr, 2024). In addition, the Deloitte Center for Integrated Research indicates that 79 % of business leaders anticipate that GenAI will catalyze substantial industry changes within the next three years, as it enhances data-driven decision-making through deeper insights and the automation of complex tasks (Deloitte, 2024). GenAI’s ability to process vast amounts of information and generate coherent and contextually relevant outputs offers the promise of improved decision accuracy and operational efficiency, given industries’ increasing reliance on data-driven insights. However, recent studies suggest that, while GenAI models, such as ChatGPT and Gemini, provide powerful generative capabilities, they may still lack the reliability required for high-stakes decision-making, particularly in academic and research contexts (Garg et al., 2024). Moravec et al. (2024) confirm that high digital literacy levels correlate with an increased propensity to utilize ChatGPT, especially for low-risk decisions, such as exploration and enjoyment. Meanwhile, the rapid adoption of GenAI has encouraged a re-examination of existing legal, ethical, and regulatory frameworks, as its ability to generate original content challenges the traditional notions of authorship, ownership, and accountability in decision-making processes (Al-Busaidi et al., 2024).

Despite the potential benefits and expanding research, the existing literature on GenAI in decision-making is highly fragmented. In recent years, numerous studies from diverse research areas have examined different aspects of GenAI applications in decision-making. Some studies have focused on the utilization of GenAI in medical diagnostics and clinical decision-making (Brügge et al., 2024; Miao et al., 2024), while others have explored its role in legal reasoning and ethical frameworks in judicial processes (Cardoso et al., 2024; Perona & de la Rosa, 2024). Other studies have addressed the impact of GenAI on business and financial decision-making as well as its potential in educational and collaborative environments (Bukar et al., 2024a; Jiang et al., 2024; Yadav et al., 2024). This disciplinary dispersion not only makes it challenging for researchers and practitioners to gain a comprehensive understanding of the field, but also raises the risk of duplicative efforts and missed opportunities for cross-disciplinary innovation. Therefore, synthesizing existing research is imperative for advancing the literature and guiding future studies to build on a consolidated knowledge base.

This study primarily aims to integrate and synthesize the diverse body of research on the role of GenAI in decision-making. Following the approach utilized by recent review articles on GenAI-related issues (Dwivedi, 2025; Dwivedi et al., 2024), this study reviews and categorizes studies from various domains to provide a comprehensive analysis that elucidates both the potential and limitations of GenAI-driven approaches in decision-making. Therefore, the study aims to clarify the current state of the field, identify knowledge gaps, and propose directions for future research. This synthesis is intended to serve as a valuable resource for scholars and practitioners in facilitating the development of informed and effective decision-making systems that leverage the strengths of GenAI.

Table 1 provides a structured overview of the literature on GenAI research in decision-making. The studies included in Table 1 were identified through a systematic search approach (Dwivedi, 2025; Dwivedi et al., 2024), using carefully chosen keywords ("Generative AI," "GenAI," "Generative Artificial Intelligence," as well as "Decision Making" and "Decision-Making") in the Scopus database. The initial search results were then manually screened to ensure that only relevant studies were included. These studies were subsequently analyzed and categorized into distinct groups, each representing a thematic focus in decision-making: medical and clinical, legal, business and financial, educational, collaborative and multidisciplinary, tourism and consumer, technological systems and frameworks, and ethical issues in decision-making. Table 1 serves as a concise yet comprehensive map of the current research landscape, offering a clear snapshot of the various domains in which GenAI impacts decision-making. In this paper, the terms GenAI and ChatGPT are used interchangeably to refer to the same underlying AI model and its applications across various contexts.

Table 1.

Generative AI research on decision-making: themes, descriptions, and citations.

Sec#  Theme  Description  Citations 
Medical and Clinical Decision-Making (41 Studies)  Focuses on how ChatGPT and other AI systems are used in healthcare, clinical decision-making, and diagnostics  Abdelwahed (2024), Beilby and Hammarberg (2024), Borna et al. (2024), Brügge et al. (2024), Camacho (2024), Chen et al. (2024b), Civettini et al. (2024), Ebrahimian et al. (2023), Griewing et al. (2023), Griewing et al. (2024), Gumilar et al. (2024), Haemmerli et al. (2023), Hager et al. (2024), Huang et al. (2023), Huo et al. (2024), Joshi et al. (2024), Ke et al. (2024), Lawson et al. (2024), Lee et al. (2023), Levartovsky et al. (2023), Lukac et al. (2023), Marchi et al. (2024), McDarby et al. (2024), Miao et al. (2024), Mykhalko et al. (2023), Nechay et al. (2024), Palenzuela et al. (2024), Rajjoub et al. (2024), Saban and Dubovi (2024), Salihu et al. (2023), Sarangi et al. (2024), Scaioli et al. (2024), Schmidl et al. (2024), Schmidt et al. (2024), Schumacher et al. (2024), Sblendorio et al. (2024), Shi et al. (2023), Teixeira-Marques et al. (2024), Ulloa et al. (2023), Zhang et al. (2023), Zhang and Liang (2024) 
Legal Decision-Making (5 Studies)  Examines the application of ChatGPT in legal decision-making contexts, including courtroom and regulatory uses  Bukar et al. (2024b), Cardoso et al. (2024), Lin and Clark (2024), Perona and de la Rosa (2024), Socol De La Osa and Remolina (2024) 
Business and Financial Decision-Making (13 Studies)  Covers the role of ChatGPT in business and financial decision-making, including supply chain, real estate, and tourism  Ahmad et al. (2024), Gloria et al. (2024), Gholami (2024), Hao et al. (2024), Jiang et al. (2024), Kim et al. (2025), (2024), Lakkaraju et al. (2023), Markova et al. (2024), Siddik et al. (2024), Stergiou and Nella (2024), Wu et al. (2024), Yadav et al. (2024) 
Educational Decision-Making (5 Studies)  Explores ChatGPT's integration into education-related decision-making, including teaching frameworks and policies  Bukar et al. (2024a), (2024b), Hiranchan and Kiattisin (2024)), Scaioli et al. (2024), Skryd and Lawrence (2024) 
Tourism and Consumer Decision-Making (5 Studies)  Investigates ChatGPT's influence on tourism and consumer behavior, particularly in autonomous travel decisions  Christensen et al. (2024), Kim et al. (2025), (2024), Stergiou and Nella (2024), Wong et al. (2023) 
Technological Systems and Frameworks (15 Studies)  Examines ChatGPT's role in enhancing technological systems, such as IoT, robotics, and real-time decision-making  Chen et al. (2024b), Ding et al. (2024), Kuang et al. (2024), Liao et al. (2023), Loya et al. (2023), Markova et al. (2024), Shan and Michel (2024), Seita and Kurahashi (2024), Shi et al. (2023), Trillo, Martínez et al. (2024), Wu et al. (2024), Xiang et al. (2024), Xu et al. (2024), Zhang et al. (2023), Zhong et al. (2023) 
Ethical Issues in Decision-Making (10 Studies)  Addresses the ethical implications, dilemmas, and challenges of using ChatGPT in various decision-making contexts  Solaiman (2024), Sel et al. (2024), Carullo (2023), Kromidha and Davison (2024), Schmidt et al. (2024), Seita and Kurahashi (2024), Malloy and Gonzalez (2024), Lin and Clark (2024), Lin et al. (2024), Goktas (2024) 
Collaborative and Multidisciplinary Decision-Making (12 Studies)  Looks at how ChatGPT supports collaborative and multidisciplinary decision-making in diverse domains  Trillo, Cabrerizo et al. (2024), Trillo, Martínez et al. (2024), Joshi et al. (2024), Shi et al. (2023), Malloy and Gonzalez (2024), Pitkäranta and Pitkäranta (2024), Ding et al. (2024), Sel et al. (2024), Schmidt et al. (2024), Huo et al. (2024), Xiang et al. (2024), Hao et al. (2024) 

Table 1 indicates that the literature on GenAI in decision-making is fragmented and lacks a consolidated framework and adequate understanding of GenAI’s impact across many domains, emphasizing the gap in addressing ethical challenges and sector-specific applications. Therefore, this study seeks to fill this gap and contribute to a better understanding of the role of GenAI in decision-making. Theoretically, it synthesizes scattered research from various fields, including medicine, law, business, and education, into a conceptually structured framework that helps researchers connect fragmented insights and identify key areas for improvement. The study findings can expand theoretical knowledge by highlighting existing patterns and gaps between disciplines, helping in developing a more holistic understanding of GenAI’s impact on decision-making. Practically, this study provides valuable insights for policymakers, professionals, and industry leaders. It identifies concrete pathways for responsible AI integration through improving AI-driven clinical diagnostics and ensuring fairness in legal rulings to enhance financial recommendations. This demonstrates the development of systems that augment human judgment while fostering transparency and trust.

The remainder of this paper is organized as follows. Section 2 deliberates the use of GenAI in diagnosis, treatment planning, and medical education, as well as its drawbacks, such as disinformation and biases. Section 3 examines the legal consequences of incorporating GenAI in court cases, including its potential to improve efficiency and address ethical and transparency concerns. Sections 4 and 5 discuss GenAI's involvement in corporate and financial and educational decision-making, respectively. Section 6 describes GenAI's impact on tourism, while sections 7 and 8 discuss technical systems and AI-driven ethical issues in decision-making. Section 9 illustrates decision-making in collaborative and multidisciplinary environments. Section 10 provides a detailed discussion and insights from the investigated domains. Section 11 outlines the domain-specific propositions and sets directions for future research on the impact of GenAI on decision-making processes. Section 12 concludes by showing the overall impact of GenAI on decision-making, benefits, challenges, and future outlook.

Medical and clinical decision-making

We found 41 recent studies on the role of GenAI in medical and clinical decision-making that have explored its applications, limitations, and ethical implications across various domains. A common argument is that ChatGPT performs well in structured decision-support tasks, particularly when summarizing medical guidelines (Miao et al., 2024), assisting with licensing and residency exams (Ebrahimian et al., 2023; Scaioli et al., 2024), and answering standardized clinical questions (Lee et al., 2023). Studies in reproductive medicine (Chen et al., 2024a) and colorectal cancer screening (Camacho, 2024) have highlighted the ability of ChatGPT in improving accessibility to medical information, helping both clinicians and patients understand treatment options. However, in areas requiring nuanced clinical reasoning, such as hematopoietic stem cell transplantation (Civettini et al., 2024) and breast cancer tumor board decisions (Griewing et al., 2023, 2024; Ulloa et al., 2023), ChatGPT has demonstrated inconsistencies and a lack of the depth required for patient-specific recommendations. These findings suggest that, while ChatGPT can serve as a valuable tool for summarizing knowledge and offering initial guidance, it remains inadequate for making autonomous medical decisions.

Multiple studies have identified the tendency of ChatGPT to generate misinformation, including fabricated references and misleading treatment recommendations, as a major limitation. Studies on hernia-related clinical decision-making (Nechay et al., 2024) and lumbar spinal stenosis treatment (Rajjoub et al., 2024; Zhang & Liang, 2024) have found that ChatGPT occasionally provided plausible yet inaccurate information, potentially misleading clinicians if not properly verified. Similar concerns were noted in uveitis-related decisions (Schumacher et al., 2024) and emergency plastic surgery management (Borna et al., 2024), where AI-generated responses sometimes conflicted with the established clinical guidelines. In addition, the issue of "hallucinations," where GenAI responses lacked coherence with current evidence-based practice, was observed in studies on surgical decision-making for gastroesophageal reflux disease (Huo et al., 2024) and gynecologic oncology (Gumilar et al., 2024). These findings reveal GenAI’s unreliability, despite its ability to provide medical insights; they also indicate the need for human oversight in the use of GenAI, particularly in high-stakes healthcare environments.

Many studies have highlighted the role of ChatGPT in medical education, particularly in training students and early-career professionals in clinical decision-making. Brügge et al. (2024) demonstrate that AI-generated structured feedback significantly improved medical students’ clinical reasoning. Similarly, studies on nursing education (Saban & Dubovi, 2024) and scoliosis treatment decision-making (Shi et al., 2023) have noted that ChatGPT helped structure clinical knowledge, making it an effective educational tool. Additionally, ChatGPT excels in standardized medical exams, often outperforming human test-takers (Lee et al., 2023; Scaioli et al., 2024). However, ChatGPT has shown limitations in scenarios requiring real-time adaptability, such as interpreting radiology images (Sarangi et al., 2024) and oncology case discussions (Lukac et al., 2023), as it lacks the ability to integrate evolving clinical data. These findings suggest that, while ChatGPT can supplement traditional medical education, it is not a substitute for experiential learning and expert mentorship.

Another challenge is ChatGPT’s susceptibility to cognitive biases in clinical decision-making. Schmidt et al. (2024) note that ChatGPT exhibited intrinsic biases, particularly when analyzing patient cases with misleading but salient features. Ke et al. (2024) explore a multi-agent framework to mitigate these biases and demonstrate that collaborative GenAI interactions improved diagnostic accuracy. In oncology, McDarby et al. (2024) and Schmidl et al. (2024) show that patient demographics influenced ChatGPT treatment recommendations, raising concerns about potential inequities in GenAI-driven medical guidance. Salihu et al. (2023) and Hager et al. (2024) identify inconsistencies in ChatGPT outputs based on prompt phrasing, further highlighting the model’s susceptibility to cognitive distortions. While AI has the potential to improve diagnostic consistency, safeguards are necessary to detect and correct biases to ensure fairness in clinical decision-making.

Several studies have emphasized the role of ChatGPT in patient-centered care and shared decision-making, particularly in increasing access to health information. Research in fertility care (Beilby & Hammarberg, 2024) and colorectal cancer screening (Camacho, 2024) has found that ChatGPT enhanced patients’ understanding of medical options, reducing information asymmetry. Similar findings emerged in neuro-oncology decision support (Lawson et al., 2024) and participatory healthcare technology design (Joshi et al., 2024), in which ChatGPT facilitated patient-clinician discussions and helped to generate patient-centered recommendations. However, concerns persist about misinformation, particularly in studies where ChatGPT’s medical advice does not align with clinical expertise, such as breast cancer management (Griewing et al., 2023; Lukac et al., 2023). Although AI can democratize medical knowledge, integration into patient care requires careful monitoring to ensure its accuracy and reliability.

Despite these limitations, researchers have highlighted the potential of ChatGPT in augmenting clinical workflows by reducing the cognitive load and improving efficiency in certain medical tasks. Studies in hypertension management (Miao et al., 2024) and radiation oncology (Huang et al., 2023) found that ChatGPT effectively summarized guidelines and assisted in treatment planning, demonstrating its value as a decision support tool. However, its inability to function autonomously, where ChatGPT’s recommendations often require human verification, has been evident in studies on acute ulcerative colitis (Levartovsky et al., 2023) and tumor board decision-making (Ulloa et al., 2023). Studies on GenAI transparency and ethical considerations (Sblendorio et al., 2024; Zhang et al., 2023) have emphasized the need for structured regulatory frameworks to ensure that AI-generated outputs align with professional standards. These findings reinforce that ChatGPT deployment must be carefully controlled to avoid over-reliance on AI-generated information in clinical workflows.

In summary, ChatGPT shows significant promise across a variety of domains, including medical decision-making, patient engagement, workflow efficiency, and improving clinical accuracy. Additionally, it supports medical education through structured feedback, helps to reduce cognitive biases in decision-making, and facilitates patient-centered care. ChatGPT’s role in summarizing guidelines and assisting in treatment planning within these domains highlights its value in the healthcare sector. However, its limitations must be addressed before broader clinical adoption. Key areas requiring improvement include GenAI transparency (Zhang et al., 2023), domain-specific training (Gumilar et al., 2024; Marchi et al., 2024), and stronger GenAI-human collaboration to enhance decision accuracy (Miao et al., 2024; Ulloa et al., 2023). Some researchers advocate the integration of ChatGPT as a supplementary tool in clinical workflows (Borna et al., 2024; Sblendorio et al., 2024), while others emphasize the necessity of regulatory oversight to mitigate the risks related to bias, misinformation, and ethical concerns (Hager et al., 2024; Salihu et al., 2023). The consensus across these studies is that ChatGPT’s capabilities should be rigorously validated and carefully implemented to enhance, rather than replace, human expertise in medical and clinical decision-making.

Legal decision-making

In legal decision-making, five recent studies have investigated the role of GenAI, focusing on its implementation in judicial systems as well as the ethical implications and regulatory frameworks regarding its use. A key theme across multiple studies is the potential of GenAI to enhance efficiency in legal decision-making despite concerns about transparency and fairness. Cardoso et al. (2024) highlight global examples, including Pakistan’s GenAI-assisted court rulings and China’s smart courts, demonstrating the growing role of GenAI in adjudication. Similarly, Perona and de la Rosa (2024) analyze a Colombian case in which a judge explicitly referenced ChatGPT in a court ruling, sparking a debate on GenAI's influence on legal reasoning. However, studies have also emphasized the risks of algorithmic bias and a lack of regulatory clarity: for instance, Socol De La Osa and Remolina (2024) argue that unregulated GenAI use in courtrooms, particularly in jurisdictions such as Colombia, Peru, and India, poses risks to judicial independence and public trust. Therefore, GenAI integration requires structured oversight to ensure fairness and adherence to legal principles.

Another key concern is the ethical implications of using ChatGPT in decision-making frameworks, particularly in balancing GenAI’s benefits with risks, such as misinformation and bias. Bukar et al. (2024) propose a structured decision-making model using the Analytic Hierarchy Process (AHP) to evaluate ethical concerns, highlighting privacy, misinformation, and academic integrity as major issues. Their findings align with those of Socol De La Osa and Remolina (2024), who advocate ex-ante and ex-post safeguards, including algorithmic fairness assessments and iterative audits, to mitigate GenAI-induced biases. Similarly, Cardoso et al. (2024) emphasize the black box problem in GenAI decision-making, in which the lack of transparency in GenAI-generated legal reasoning undermines accountability. The ethical use of ChatGPT in legal decision-making necessitates proactive regulatory mechanisms such as transparency mandates and human oversight to prevent unintended legal and societal consequences.

While most studies focus on GenAI’s impact on judicial decision-making, Lin and Clark (2024) extend the discussion to GenAI’s role in ethical reasoning and public decision-making beyond the courtroom. Their research on how first-year public-speaking students navigate ethical dilemmas when using ChatGPT highlights broader concerns regarding moral intensity, ethical recognition, and behavioral intentions in AI applications. These findings relate to the concerns raised by Perona and de la Rosa (2024) regarding the ethical risks of over-reliance on ChatGPT in judicial settings. The overlap between legal and educational contexts suggests that GenAI’s ethical and regulatory challenges are not confined to courtrooms, but extend to broader decision-making frameworks that shape public discourse and institutional governance. Overall, there is a need for comprehensive GenAI policies that not only address the ethical concerns and regulatory gaps identified across various studies but also ensure the responsible integration of AI tools into decision-making frameworks. This includes the development of transparent, accountable, and fair systems that protect the rights of individuals while promoting the efficient and accurate use of AI in complex contexts such as legal decision-making. Additionally, continuous oversight and reassessment are necessary, as AI technologies evolve to ensure that they align with changing societal and legal expectations.

Business and financial decision-making

Recently, 13 studies on the role of GenAI in business and financial decision-making have illustrated its applications across industries, from investment analysis and supply chain resilience to sustainability and financial advisement. A common theme across multiple studies is the ability of GenAI-driven decision systems to enhance efficiency and optimize business operations. Jiang et al. (2024) and Yadav et al. (2024) demonstrate how GenAI-powered financial and knowledge systems improved data retrieval, filtering, and classification, aiding users in decision-making through sentiment analysis and structured financial insights. Similarly, Gloria et al. (2024) show that domain-specific GenAI models, such as Real-GPT, can enhance real estate investment decisions by outperforming general-purpose models in market analysis. However, while AI enhances data processing and operational decision-making, Lakkaraju et al. (2023) raise concerns about fairness and bias in GenAI-driven financial advisement, arguing that inconsistencies in recommendations across user groups pose ethical and regulatory risks. Although GenAI improves efficiency and accuracy in financial decision-making, mitigating biases and ensuring transparency continue to present critical challenges.

Another major consideration is the impact of GenAI on sustainability and industrial decision-making. Wu et al. (2024) and Gholami (2024) explore GenAI-driven frameworks designed to optimize carbon emission management and sustainable manufacturing and demonstrate how GenAI models improve efficiency through event logs and historical emissions data analyses. Markova et al. (2024) expand this by proposing on-premise GenAI frameworks for real-time industrial automation, highlighting the benefits of local GenAI deployment for predictive maintenance and process optimization. These studies suggest that GenAI contributes significantly to sustainability goals by enhancing operational adaptability and reducing resource consumption. However, challenges such as data security concerns (Markova et al., 2024) and computational costs (Wu et al., 2024) underscore the need for scalable and cost-effective GenAI solutions that align with industry-specific constraints. Collectively, these studies indicate that GenAI has the potential to facilitate sustainable business practices; however, successful implementation requires balancing efficiency gains with cost, security, and compliance considerations.

The role of GenAI in business resilience and risk management is another prominent research area. Ahmad et al. (2024) propose a GenAI-enhanced decision-making framework for supply chain resilience, arguing that GenAI’s ability to anticipate risks and optimize mitigation strategies strengthens SMEs' responses to disruptions. Similarly, Hao et al. (2024) investigate the intersection of human intelligence and GenAI in collaborative business decision-making and found that GenAI reduces cognitive biases and enhances adaptability. By contrast, Kim et al. (2025), (2024) focus on consumer-facing applications, examining how misinformation and AI errors affect trust and decision-making in the financial and tourism sectors. Their findings suggest that AI aids in business continuity and risk mitigation, improving resilience in dynamic business environments. However, it can introduce new vulnerabilities, such as algorithmic biases or misinformation, which may erode user confidence. Therefore, human oversight and structured governance are necessary to maintain trust and mitigate unintended risks.

A final consideration is the influence of GenAI on funding success and market dynamics. Siddik et al. (2024) highlight that investor influence, rather than technological factors, significantly determines funding outcomes for GenAI startups, suggesting that venture capital networks play a more critical role in financial success than GenAI capabilities alone. By contrast, Jiang et al. (2024) and Wu et al. (2024) emphasize the growing importance of GenAI in financial forecasting and sustainability decision-making, indicating that GenAI-driven insights shape investment decisions and market trends. Evidently, GenAI is a valuable tool for financial analysis, which enhances market insights and operational decision-making. This divergence in findings suggests that investor confidence and external economic factors remain dominant in shaping financial success, as broader financial ecosystems, regulatory considerations, and human judgment continue to shape business outcomes.

Educational decision-making

We reviewed five recent studies that examined the use of GenAI in educational decision-making, focusing on policy formation, instructional assistance, and student learning, alongside ethical and practical challenges. A recurring theme across multiple studies is the need for structured frameworks to guide GenAI integration in education. Bukar et al. (2024) propose a Risk, Reward, and Resilience (RRR) model to balance GenAI’s benefits, such as personalized learning, with the risks being misinformation and academic dishonesty, among others. Furthermore, Bukar et al. (2024b) utilize the Analytic Hierarchy Process (AHP) to prioritize ethical concerns, emphasizing that policies must balance restrictions with regulations to maintain academic integrity. These findings align with those of Hiranchan and Kiattisin (2024)), who advocate structured prompt frameworks to improve the accuracy of GenAI-driven decision-making, particularly in instructional settings. Collectively, these studies indicate that, while GenAI can improve efficiency and personalization in education, its integration must be guided by policies that address ethical concerns and ensure equitable access.

Importantly, the use of GenAI in medical education emphasizes its benefits and drawbacks in clinical training. Skryd and Lawrence (2024) examine ChatGPT’s use in ward-based teaching and found that it effectively supports knowledge reinforcement and team discussions, but struggles in high-stakes, emergency scenarios requiring nuanced reasoning. Similarly, Scaioli et al. (2024) evaluate ChatGPT’s performance on medical residency exams in Italy, demonstrating its strong analytical capabilities, while cautioning against over-reliance owing to its inability to process recent medical advancements. These studies collectively underscore AI function as an effective supplementary educational tool, particularly for reinforcing theoretical knowledge. However, they highlight AI’s limitations in complex decision-making scenarios that require clinical judgment and real-world adaptability and underline the need for structured guidelines and human oversight to ensure the responsible use of GenAI in medical training and assessment.

Despite the potential of GenAI in education, ethical challenges and implementation concerns remain common across the studies. Bukar et al. (2024a) and (2024b) emphasize developing adaptive regulations that can evolve alongside GenAI advancements to prevent misuse and academic inequities. Similarly, Skryd and Lawrence (2024) highlight misinformation risks and biases in GenAI-generated medical responses, reinforcing the need for ethical guidelines in GenAI-driven instruction. Hiranchan and Kiattisin (2024)) add to this discussion by submitting that GenAI effectiveness depends on prompt structuring, which can influence bias, accuracy, and decision relevance. In summary, although GenAI can improve educational decision-making, its efficacy depends not only on performance but also on established frameworks and ethical principles that regulate its application. The successful incorporation of GenAI in education necessitates regulations that harmonize innovation with responsible governance and safeguard personalized learning and academic integrity. These studies illustrate that frameworks such as the RRR model and AHP are essential for resolving ethical issues and directing the proper use of GenAI in diverse educational contexts.

Tourism and consumer decision-making

ChatGPT’s ability to enhance the efficiency and personalization of travel decisions is a recurring theme in the five recent studies that analyzed how GenAI affects trip planning, trust in AI recommendations, and misinformation concerns in tourism and consumer decision-making. Wong et al. (2023) highlight how ChatGPT improves decision-making across all travel phases (pre-trip, en-route, and post-trip) by offering customized itineraries, real-time assistance, and content creation. Similarly, Stergiou and Nella (2024) emphasize ChatGPT’s capacity to tailor recommendations based on individual preferences, demonstrating how GenAI enhances travel information accessibility and diagnosticity. However, Kim et al. (2024) caution that trust in GenAI recommendations is contingent on user experience, because exposure to GenAI errors can diminish adoption intention. Although ChatGPT streamlines travel planning and enhances consumer experiences, ensuring reliability and transparency is critical in maintaining user confidence.

The issue of GenAI-generated misinformation presents a significant concern in consumer tourism decision-making. Christensen et al. (2024) and Kim et al. (2025) investigate the impact of GenAI hallucinations, in which ChatGPT produces plausible but inaccurate information, on travelers’ perceptions and behaviors. Christensen et al. (2024) find that, while consumers recognize the risk of GenAI misinformation, they often continue to prefer GenAI-generated itineraries due to perceived impartiality and personalization. Conversely, Kim et al. (2025) demonstrate that incorrect information significantly erodes trust, particularly when it is domain-specific and dominant within GenAI-generated recommendations. These findings highlight a paradox in which GenAI’s convenience and customization drive its adoption despite the impact of misinformation on consumer confidence. Both studies have underscored the need for user education and enhanced transparency measures in mitigating GenAI-induced errors in travel planning.

Despite these challenges, research has suggested that GenAI’s role in tourism decision-making will continue to evolve, provided that its limitations are effectively managed. Kim et al. (2024) and Stergiou and Nella (2024) argue that the effectiveness of ChatGPT depends on balancing user expectations with system reliability. While Kim et al. (2024) report that positive information about ChatGPT increases adoption intention, Wong et al. (2023) emphasize that regulatory oversight and technological advancements are necessary in addressing biases, privacy concerns, and outdated knowledge bases. Therefore, ChatGPT’s success in tourism depends on continuous improvements in AI accuracy, user education, and ethical GenAI deployment to ensure consumers make informed and confident travel decisions. For instance, transparency regulations that require AI platforms to provide explicit disclosures regarding the generation of recommendations enable users to comprehend their limits more effectively. In addition, bias-mitigation techniques must be included in GenAI systems to detect and rectify biases in real time, guaranteeing that recommendations are equitable and fair.

Technological systems and frameworks

Optimization of GenAI-driven decision systems to improve efficiency, accuracy, and adaptability in real-time applications is a central theme across 15 recent studies that have investigated the integration of GenAI into industrial automation, robotics, autonomous decision-making, and emergency response. Ding et al. (2024) and Xiang et al. (2024) demonstrate how incorporating multi-modal data integration and reward-shaping mechanisms enables GenAI-enhanced reinforcement learning frameworks to improve decision-making in complex environments, such as the metaverse and Unmanned Air Vehicle air combat simulations. Similarly, Markova et al. (2024) and Wu et al. (2024) highlight the benefits of integrating ChatGPT-driven models into Industry 4.0 systems, particularly for predictive maintenance and carbon emissions management. These studies indicate that GenAI plays an increasingly important role in optimizing high-stakes and real-time decision-making across multiple sectors; however, challenges such as computational demands and system scalability remain areas for further refinement.

Another significant focus is GenAI’s role in knowledge extraction and domain-specific decision frameworks, particularly in improving system reliability and interpretability. Xu et al. (2024) and Zhong et al. (2023) explore how ChatGPT and other large language models (LLMs) can support engineering design and tax-related decision-making to assist in extracting technical knowledge from diverse sources, including science fiction narratives and structured databases. Chen et al. (2024) extend this discussion by integrating knowledge graphs with LLMs to improve emergency response decision-making and ensure that GenAI-generated recommendations adhere to regulatory guidelines. GenAI’s effectiveness in decision-making is enhanced when paired with structured knowledge systems, reducing hallucinations and improving the accuracy of GenAI-generated insights. However, Xu et al. (2024) caution that, while GenAI can efficiently retrieve factual knowledge, its subjective reasoning still lags behind that of human decision-makers, underscoring the need for hybrid GenAI-human collaboration in complex judgment tasks.

Several studies have investigated the integration of GenAI in real-time automation and robotics, in which the combination of LLMs with physical systems is key in improving autonomous functionality. By integrating language processing with visual perception, Liao et al. (2023) and Kuang et al. (2024) explore how ChatGPT-driven decision frameworks enhance robotic grasping tasks and traffic scene comprehension, respectively. Shan and Michel (2024) apply a similar hybrid approach to gaming, combining goal-oriented action planning with ChatGPT to optimize real-time strategic decision-making. These studies suggest that GenAI’s ability to process unstructured data in real time enhances system adaptability across different environments, from industrial robotics to interactive digital applications. However, they highlight technical limitations, such as latency issues (Shan & Michel, 2024) and reduced performance in complex environments with unpredictable conditions (Kuang et al., 2024), underscoring the need for further optimization of GenAI-driven automation.

Finally, multiple studies address the role of GenAI in improving the interpretability and ethical alignment of technological decision systems. Trillo, Martinez et al. (2024) and Seita and Kurahashi (2024) explore the impact of ChatGPT on group decision-making by enhancing consensus-building and reducing bias in decision outputs. Similarly, Zhang et al. (2023) propose a framework that integrates Markov logic networks with LLMs to improve transparency in GenAI-driven medical diagnostics, ensuring that GenAI-generated insights align with evidence-based practices. Although these studies highlight GenAI’s potential to enhance structured decision-making, they also stress the importance of implementing safeguards to prevent over-reliance on GenAI-generated conclusions. Research on technological systems and frameworks suggests that GenAI’s decision-making capabilities are most effective when paired with structured rule-based models, domain-specific adaptations, and continuous human oversight to ensure accuracy, efficiency, and ethical integrity.

Ethical decision-making

GenAI’s effects on transparency, bias, accountability, regulatory challenges, and ethical decision-making across various domains have been addressed in 10 recent studies. A recurring theme is the need for structured oversight and governance to mitigate the risks associated with GenAI-driven decisions. Solaiman (2024) and Carullo (2023) emphasize the lack of clear regulatory safeguards in mental health and public administration, in which GenAI-driven decisions can have serious consequences due to biased outputs and a lack of transparency. Similarly, Lin et al. (2024) assess ChatGPT's performance on ethical judgment tests and find that, although GenAI models, such as GPT-4, can demonstrate high-level ethical reasoning, they lack the accountability required for professional decision-making. Furthermore, Goktas (2024) highlights that challenges related to transparency and explainability in GenAI-driven decision systems have become increasingly prominent, particularly in healthcare and finance, necessitating continuous monitoring and sector-specific ethical guidelines. These studies argue for GenAI deployment guided by robust ethical and regulatory frameworks to prevent unforeseen consequences.

Several studies have explored the significant issue of bias in GenAI-driven decision-making. Schmidt et al. (2024) investigate bias in GenAI-assisted diagnostic decision-making, finding that ChatGPT exhibits intrinsic biases similar to those of human professionals, although it remains unaffected by contextual extrinsic biases. Similarly, Sel et al. (2024) propose a Skin-in-the-Game (SKIG) framework to improve GenAI moral reasoning by incorporating stakeholder perspectives, thereby demonstrating improvements in ethical decision accuracy. Meanwhile, Kromidha and Davison (2024) assess the limitations of GenAI in business decision-making, arguing that GenAI lacks the contextual judgment necessary for fully autonomous ethical choices. Although GenAI can process and structure ethical reasoning, it struggles with context sensitivity, bias mitigation, and moral accountability. Consequently, GenAI decision-making must be supplemented with human oversight and structured interventions such as bias-detection frameworks and ethical alignment strategies to ensure fairness and reliability.

A final area of discussion is the role of GenAI in group-based ethical decision-making and its implications for replacing or supplementing human judgment. Seita and Kurahashi (2024) find that ChatGPT's decision-making closely mirrors human group discussions, raising questions about whether GenAI could serve as an alternative to resource-intensive deliberations. However, Malloy and Gonzalez (2024) argue that GenAI-driven cognitive models require improvements in generalization and predictive accuracy, particularly in high-stakes ethical dilemmas. Lin and Clark (2024) extend this discussion to academic ethics by analyzing how students perceive GenAI-generated content in plagiarism detection and public speaking contexts. Thus, GenAI provides structured reasoning and facilitates ethical decision-making processes. However, its ethical limitations raise concerns about accountability in educational and professional assessments, and thus, its application should be limited to augmenting human judgment rather than replacing it, ensuring that GenAI remains a tool for structured deliberation rather than an autonomous moral authority. Although the ethical limitations of GenAI are predominantly articulated from a Western perspective, ethical frameworks and cultural norms vary across countries. In several non-Western environments, decision-making procedures may incorporate community-based or hierarchical frameworks, influencing the perception and implementation of GenAI's role in ethical decision-making.

Collaborative and multidisciplinary decision-making

We uncovered 12 recent studies on GenAI in collaborative and multidisciplinary decision-making that address its potential to improve consensus building, cognitive modeling, and complex decision environments. The majority of the studies explore GenAI’s ability to streamline group decision-making processes through structuring discussions and mitigating biases. Trillo, Cabrerizo et al. (2024) and Trillo, Martínez et al. (2024) explore how ChatGPT and other LLMs assist in group decision-making by analyzing sentiments in discussions, reducing the influence of hostility, and adjusting the weight assigned to each expert to foster constructive debates. Similarly, Joshi et al. (2024) demonstrate how AI-driven participatory design tools improve collaborative decision-making in healthcare technology development by enabling non-technical stakeholders to contribute meaningfully. By contrast, Schmidt et al. (2024) examine GenAI’s susceptibility to intrinsic biases in multidisciplinary group decisions, noting that ChatGPT remains prone to cognitive distortions, similar to human experts. These studies suggest that careful oversight over GenAI is required to manage its biases and ensure equitable stakeholder participation.

Moreover, GenAI integration into collaborative environments that require real-time decision support has been highlighted. Xiang et al. (2024) propose a GenAI-enhanced reinforcement learning system that improves Unmanned Air Vehicle air combat decision-making by incorporating LLMs for strategic adjustments. Similarly, Ding et al. (2024) introduce a multi-dimensional optimization framework for metaverse-based decision-making, demonstrating how GenAI can facilitate real-time interactions and autonomous adaptation in dynamic digital environments. Pitkäranta and Pitkäranta (2024) extend these findings to corporate decision-making and develop a RAGADA framework to align GenAI-driven insights with organizational objectives and ethical standards. These studies emphasize GenAI’s ability to enhance decision efficiency and adaptability through GenAI-driven decision systems that improve real-time collaboration across disciplines. Conversely, they highlight the challenges related to computational costs, ethical compliance, and model retraining, and that transparency and reliability remain ongoing concerns.

The role of GenAI in enhancing cognitive modeling and teamwork is another key theme in the literature. Malloy and Gonzalez (2024) propose a generative memory framework to improve cognitive modeling in decision science and found that GenAI-enhanced memory representation and action prediction can improve learning efficiency. Similarly, Hao et al. (2024) investigate human-AI collaboration in organizational decision-making, revealing that GenAI reduces cognitive biases and enhances analytical reasoning, but also increases the risk of over-reliance on AI-generated insights. Shi et al. (2023) explore a related concept in medical decision-making by developing Chat-Orthopedist, a retrieval-augmented GenAI model that supports shared decision-making by integrating structured medical knowledge with ChatGPT. Although GenAI enhances cognitive modeling and structured decision support, human expertise should be balanced with GenAI recommendations to maintain decision quality and prevent automation bias.

Furthermore, ethical considerations and GenAI’s influence on stakeholder alignment refine the discussion on GenAI-driven collaborative decision-making. Sel et al. (2024) introduce the SKIG framework to improve GenAI’s moral reasoning by simulating accountability and stakeholder perspectives, thereby demonstrating its effectiveness in addressing ethical dilemmas. Using sentiment analysis and preference weighting to enhance group coherence, Huo et al. (2024) and Trillo, Martínez et al. (2024) similarly emphasize GenAI’s role in ensuring fairness in multi-stakeholder environments. However, these studies also highlight GenAI’s potential to introduce unintended biases, requiring governance mechanisms to ensure transparency and fairness. GenAI success in facilitating ethical decision-making by structuring stakeholder perspectives depends on well-defined oversight frameworks that prioritize accountability and trust in human-AI collaboration. Sophisticated frameworks, such as RAGADA and SKIG, are promising but would require more empirical validation and testing to confirm their scalability and applicability across other domains. Further research should explore their practical efficacy and address any constraints on applying these frameworks to diverse decision-making contexts to guarantee their widespread application and significant outcomes.

Discussion and insights

Table 2 provides a structured comparison of how GenAI impacts decision-making across the medical, legal, business, education, collaborative, tourism, technological, and ethical decision-making domains. It highlights common themes, such as efficiency, bias, trust, and oversight, while also identifying domain-specific challenges, such as high-stakes risks in healthcare and law, personalization concerns in business and education, and automation biases in technology. Understanding these intersections helps in developing more ethical, transparent, and effective AI applications in various fields.

Table 2.

Comparative analysis of key themes in GenAI-driven decision-making across domains.

Theme  Medical  Legal  Business & Finance  Education  Collaboration  Tourism  Technology  Ethical Decision-Making 
Efficiency  Enhances diagnosis & medical training  Aids legal research & case law retrieval  Optimizes financial analysis, risk & supply chains  Automates grading, personalized learning  Reduces cognitive load in group decisions  Speeds up travel planning & real-time recommendations  Optimizes robotics & emergency response  Improves decision frameworks but raises risks 
Bias & Fairness  Demographic biases in medical decisions  Risk of systemic bias in court rulings  Unequal investment insights  Academic integrity concerns  Bias in AI-driven group discussions  Limited but present in market segmentation  Bias in automation systems  Ethical dilemmas around bias and misinformation 
Autonomy vs. Oversight  Needs human verification for medical accuracy  GenAI legal decisions require transparency & human checks  More widely accepted for financial modeling  Effective but needs structured prompts  Supports collaboration, not independent decisions  Effective but needs user verification  Automation possible, but human input is crucial  Requires strict oversight to align with values 
Trust & Misinformation  Misleading recommendations can be harmful  AI decisions lack full transparency  Trust affected by inconsistencies in AI-driven advice  Risk of misinformation in academic settings  Helps structure discussions but can be manipulated  AI errors affect consumer trust  AI models lack explainability  Needs accountability & transparency mandates 
Collaboration Potential  Assists but doesn’t replace medical professionals  Supports lawyers but lacks legal reasoning skills  Enhances business intelligence  Improves student engagement  Enhances inclusivity & structured decision-making  Personalizes travel experiences  Supports real-time industry decision-making  Can structure ethical debates but lacks moral agency 

Based on Table 2, the Venn Diagram (Fig. 1) illustrates the overlapping and unique challenges of GenAI in medical and legal, business and education, technology, and ethical decision-making.

Fig. 1.

Intersection of GenAI in decision-making across key domains.

(0.15MB).

This Venn Diagram (Fig. 1) provides a visual representation of the shared and unique challenges in GenAI decision-making across multiple sectors, reinforcing the key themes discussed in Table 2. It demonstrates that all domains benefit from efficiency gains, but share concerns about bias, transparency, and ethical dilemmas. These intersections highlight domain-specific challenges such as high oversight requirements in the medical and legal fields, personalization risks in business and education, and automation challenges in technology and ethics. Understanding these overlaps in GenAI’s roles across different domains allows for a more detailed breakdown of specific categories and their interconnections, as follows:

  • Three Major Categories:

    • Medical & Legal → High oversight and ethical concerns

    • Business & Education → Personalization and misinformation risks

    • Technology & Ethics → Bias and automation challenges

  • Overlapping Areas:

    • Medical & Legal + Business & Education → AI-assisted decision-making

    • Medical & Legal + Technology & Ethics → Transparency and trust issues

    • Business & Education + Technology & Ethics → Regulatory and governance needs

    • All Three Domains (Center Overlap) → Efficiency, bias, and ethical dilemmas

At a higher level, an evident similarity across the domains is the recognition of GenAI's ability to rapidly process large volumes of information, thereby improving decision-making efficiency. For instance, in medical decision-making, ChatGPT supports clinicians by summarizing medical guidelines and assisting with licensing exams (Ebrahimian et al., 2023; Miao et al., 2024), similar to its role in legal contexts in which it aids in case law retrieval and regulatory compliance (Cardoso et al., 2024). Similarly, in the business and financial domain, GenAI enhances data retrieval, investment analysis, and risk assessment (Jiang et al., 2024; Yadav et al., 2024). The efficiency of GenAI-driven decision support is a unifying theme across these sectors.

However, a major contrast lies in the stakes involved in different decision-making scenarios. In medical and legal contexts, GenAI errors can have profound consequences, such as misdiagnoses (Griewing et al., 2023) or biased legal verdicts (Socol De La Osa & Remolina, 2024). Meanwhile, in business and tourism, errors tend to be less severe but still consequential. GenAI-generated misinformation in tourism recommendations (Christensen et al., 2024) or in financial decision-making (Lakkaraju et al., 2023) may not have life-altering implications, such as healthcare or law, but can lead to monetary losses or poor consumer experiences.

A related contrast appears in the extent to which GenAI can operate autonomously. In business and financial decision-making, GenAI is widely accepted as a decision support tool, especially in supply chain optimization and sustainability initiatives (Ahmad et al., 2024; Gholami, 2024). However, in medicine, although GenAI assists in training medical professionals and structuring clinical knowledge, it remains unreliable for direct patient recommendations because of the risks of bias and misinformation (Shi et al., 2023; Zhang et al., 2023). Similarly, in legal settings, GenAI cannot function as an independent decision-maker because of the necessity for judicial interpretation and ethical considerations (Lin & Clark, 2024).

Another recurring theme concerns biases in GenAI-driven decision-making. In medical applications, GenAI-generated recommendations sometimes reflect demographic bias, leading to inequitable treatment options (McDarby et al., 2024; Schmidl et al., 2024). These biases manifest in business and financial contexts as inconsistencies in financial advisement, which can disproportionately affect certain user groups (Lakkaraju et al., 2023). Similarly, in legal decision-making, concerns regarding GenAI reinforcing systemic biases in judicial rulings have been raised (Socol De La Osa & Remolina, 2024). By contrast, the impact of the biases in tourism and consumer decision-making is more related to market segmentation and personalization than to structural inequities (Kim et al., 2024).

At the same time, the role of GenAI in fostering collaboration is significantly different. In collaborative and multidisciplinary decision-making, GenAI is viewed as a facilitator, structuring discussions and enhancing stakeholder engagement (Joshi et al., 2024; Trillo, Cabrerizo et al., 2024), whereas in ethical decision-making, GenAI is more of a challenge to existing norms than a facilitator, as it raises questions about transparency, accountability, and regulatory compliance (Kromidha & Davison, 2024; Solaiman, 2024). Specifically, in collaborative decision-making, GenAI is seen as a positive force for inclusivity, whereas in ethical contexts, its limitations demand scrutiny and caution.

The technological domain significantly overlaps with business and multidisciplinary decision-making but introduces additional complexities regarding GenAI’s real-time adaptability and computational constraints. Studies have highlighted the role of GenAI in enhancing Industry 4.0 systems, robotics, and emergency response decision-making (Kuang et al., 2024; Markova et al., 2024), indicating a greater reliance on GenAI automation compared to other fields. However, this sector faces scalability issues because GenAI-driven frameworks often require high computational power and lack interpretability in decision-making (Xu et al., 2024; Zhong et al., 2023).

Finally, the ethical concerns surrounding GenAI are universal but manifest differently across domains. In law and medicine, ethical considerations revolve around fairness, misinformation, and accountability (Lin et al., 2024; Schmidt et al., 2024), whereas in business and tourism, ethical concerns relate to transparency and consumer trust (Christensen et al., 2024; Kim et al., 2025). The technological domain faces ethical dilemmas concerning GenAI bias in automation and predictive decision-making (Seita & Kurahashi, 2024), whereas in education, GenAI ethics focus on misinformation risks and academic integrity (Bukar et al., 2024a, 2024b). These variations highlight the domain-specific nature of GenAI’s ethical challenges.

Overall, the reviewed studies spanning diverse fields indicate that, although GenAI possesses considerable potential to enhance decision-making efficiency and customization, its implementation must be meticulously managed to mitigate domain-specific hazards. The key pillars, including transparency mandates, risk tolerance standards, and ethical AI within stakeholder frameworks, underscore the need for thorough monitoring and regulatory systems to guarantee the responsible implementation of GenAI. To effectively harness its advantages, GenAI must be supported by human supervision, transparency protocols, and ethical safeguards to guarantee equity, precision, and trust across all domains.

Practical implications and future research directions

Based on this study’s analysis and discussion, Fig. 2 provides a structured overview of the research agenda for GenAI-driven decision-making across various domains. It depicts key propositions in the eight areas: healthcare, legal, finance, education, collaboration, travel, automation, and ethics, each of which addresses a specific challenge and potential solution. The propositions highlight critical aspects, such as improving GenAI reliability through multi-agent frameworks, enhancing transparency in legal GenAI applications, personalizing financial advisement, and implementing structured prompt engineering in education. Additionally, the research agenda emphasizes the importance of sentiment analysis for group decision-making, cross-referencing AI-generated travel recommendations, hybrid GenAI-human models for automation, and ethical oversight frameworks. The propositions depicted in Fig. 2 are discussed and formulated in the remainder of this section.

Fig. 2.

Propositions across various domains.

(0.26MB).

The integration of multi-agent AI frameworks can significantly enhance the reliability of ChatGPT in medical decision-making by reducing biases and improving diagnostic accuracy. Existing studies have highlighted the strengths of ChatGPT in summarizing medical guidelines and assisting in clinical education (Brügge et al., 2024; Miao et al., 2024). However, ChatGPT’s susceptibility to cognitive biases and misinformation remains a major limitation (Schmidl et al., 2024). Multi-agent AI frameworks, in which multiple AI models engage in collaborative analysis before generating a final recommendation, have been shown to mitigate biases in decision-making (Ke et al., 2024). By incorporating different AI perspectives and cross-referencing data, multi-agent approaches can improve AI-assisted diagnostics and minimize errors. This is especially relevant for high-stakes medical decisions in which GenAI hallucinations and biased outputs can lead to harmful outcomes. Future research should explore how multi-agent frameworks can be refined for various medical applications to ensure that GenAI-driven decisions align with clinical best practices and ethical standards.

Proposition 1

The integration of multi-agent AI frameworks can improve ChatGPT’s clinical decision-making by mitigating biases and enhancing diagnostic accuracy.

The growing role of GenAI in legal decision-making has introduced efficiency benefits, as well as significant concerns about transparency and accountability. One of the most pressing challenges is the black box nature of GenAI-generated legal reasoning, where GenAI’s decision-making process is not fully explainable (Cardoso et al., 2024). This lack of transparency undermines public trust in GenAI-assisted legal rulings and creates ethical dilemmas regarding fairness and bias (Socol De La Osa & Remolina, 2024). Implementing GenAI transparency mandates, such as requiring AI-generated rulings to include explainable reasoning, sources, and audit trails, could improve legal decision-making by ensuring AI outputs remain interpretable and accountable. Regulatory bodies should establish mechanisms for reviewing GenAI-generated legal recommendations to prevent misinformation and bias from influencing judicial decisions. Such transparency measures would not only enhance public confidence in GenAI-assisted legal processes but also provide legal professionals with reliable tools to complement rather than replace human judgment.

Proposition 2

Implementing GenAI transparency mandates can enhance public trust in GenAI-assisted legal rulings.

AI-driven financial advisement has revolutionized decision-making by providing real-time insights and predictive analytics. However, existing research indicates that GenAI-generated financial recommendations can exhibit inconsistencies and biases, which can negatively impact user groups with different investment profiles (Lakkaraju et al., 2023). To address this issue, AI-powered financial advisory systems should incorporate user-specific risk tolerance models that enable personalized decision-making. By tailoring GenAI-driven recommendations to individual investors’ risk appetites, these models can enhance decision accuracy while reducing exposure to biased outputs. Furthermore, incorporating real-time market fluctuations and sentiment analysis can improve the relevance of GenAI-generated financial guidance. This approach aligns with the growing trend of GenAI-driven personalization in financial technology, ensuring that GenAI serves as a reliable and adaptive tool for investment planning. Future research should focus on refining these models to balance the efficiency of GenAI with user-specific financial objectives.

Proposition 3

GenAI-driven financial advisement should incorporate user-specific risk tolerance models to enhance decision personalization.

The effectiveness of GenAI in education depends largely on how the prompts are structured and how GenAI-generated content is interpreted. Studies have observed that poorly designed prompts can lead to biased, misleading, or incomplete educational responses (Hiranchan & Kiattisin, 2024). To optimize GenAI’s role in instructional decision-making, educational institutions should implement structured, prompt engineering frameworks that guide GenAI interactions to produce accurate, unbiased, and pedagogically sound outputs. These frameworks include standardized question formats, cross-referencing of GenAI responses with verified sources, and integrating GenAI feedback mechanisms into learning management systems. Furthermore, structured prompt engineering can enhance GenAI’s ability to adapt to diverse learning needs, making it a more equitable educational tool. Educators should continuously review GenAI-generated content to ensure alignment with curriculum standards and prevent misinformation.

Proposition 4

GenAI-assisted educational decision-making should be guided by structured, prompt engineering frameworks to enhance accuracy and fairness.

ChatGPT and similar GenAI tools have significantly improved travel planning by offering personalized recommendations and real-time assistance (Wong et al., 2023). However, studies have underscored the risks of GenAI-generated misinformation in which users rely on inaccurate or outdated information for travel decisions (Christensen et al., 2024; Kim et al., 2025). To address this issue, GenAI-generated travel recommendations should be cross-referenced with verified sources such as government tourism boards, official hotel websites, and real-time weather and safety data. Implementing automatic verification mechanisms within GenAI-driven travel planning tools can enhance the reliability of recommendations, fostering consumer trust. GenAI models should also be updated regularly to incorporate changes in travel regulations and geopolitical developments.

Proposition 5

GenAI-generated travel recommendations should be cross-referenced with verified sources to enhance trust and accuracy.

GenAI has become a crucial component in automation and real-time decision-making; however, its adaptability remains a significant challenge. In sectors such as robotics, industrial automation, and emergency response, GenAI-driven systems functioning should be reliable in dynamic environments (Kuang et al., 2024; Markova et al., 2024). Hybrid GenAI-human decision-making models offer a promising solution by combining the computational efficiency of GenAI with human adaptability. These models allow human operators to intervene when GenAI systems encounter unfamiliar scenarios or fail to effectively process real-time data. Such an approach improves decision accuracy while ensuring that GenAI does not operate autonomously without oversight. Future research should explore optimal configurations for GenAI-human collaboration to maximize efficiency in real-time applications.

Proposition 6

Hybrid GenAI-human decision-making models can enhance the adaptability of GenAI to real-time automation tasks.

Ethical concerns surrounding GenAI-driven decision-making often stem from bias, lack of accountability, and transparency issues (Goktas, 2024; Solaiman, 2024). These concerns can be addressed by incorporating stakeholder-aligned ethical frameworks into GenAI decision-making systems. For instance, SKIG framework aligns GenAI-generated decisions with stakeholder values, improving fairness and accountability (Sel et al., 2024). By integrating similar ethical oversight mechanisms, GenAI models can minimize the risk of unintended consequences in high-stakes decisions. Policymakers and industry leaders should collaborate to establish ethical GenAI guidelines that promote transparency and equitable outcomes across sectors.

Proposition 7

Ethical GenAI decision-making should incorporate stakeholder-aligned frameworks to ensure fairness and accountability.

Effective decision-making in collaborative environments often requires balancing diverse perspectives while mitigating biases and conflicts. GenAI-enhanced sentiment analysis can play a crucial role in facilitating constructive discussions by identifying and addressing hostility in group decision-making processes (Trillo, Cabrerizo et al., 2024). Current research indicates that AI’s ability to detect sentiments and adjust discourse can significantly reduce groupthink and polarization, thereby improving consensus building (Joshi et al., 2024). By integrating GenAI-driven sentiment analysis into collaborative platforms, decision-makers can create more inclusive environments that prioritize evidence-based discussions. This approach can be particularly beneficial in corporate, medical, and policy-driven decision-making settings in which stakeholder alignment is essential. However, the effectiveness of GenAI in moderating discussions must be continuously evaluated to prevent algorithmic biases from influencing outcomes.

Proposition 8

GenAI-enhanced sentiment analysis can improve group decision-making by mitigating hostility and fostering constructive discussions.

Conclusion

GenAI has transformed decision-making across multiple sectors; however, the current research landscape remains fragmented. This literature analysis reveals that GenAI’s applications span medical diagnostics and legal adjudication to business analytics and educational frameworks, each demonstrating benefits such as enhanced efficiency, rapid data processing, and improved operational accuracy. However, these advancements are tempered by significant challenges, including pervasive biases, the risk of misinformation, and the need for robust regulatory oversight. The divergent results across various domains highlight the necessity for integrative research that consolidates scattered insights into a unified framework capable of harnessing the potential of GenAI while mitigating its inherent risks. As decision-making processes become increasingly data-driven and GenAI-augmented, maintaining a balance between technological innovation and ethical accountability is paramount. We highlight several issues to guide future research in this domain: How can future advances in GenAI redefine the core principles of decision-making in high-stakes environments? What innovative governance mechanisms can be established to ensure that GenAI’s influence on decision-making remains transparent, fair, and resilient against bias and misinformation? In sum, this work contributes to the growing body of knowledge by demonstrating how GenAI transforms decision-making through enhanced decision accuracy, greater efficiency, and personalized solutions. For instance, GenAI streamlines complicated decision-making processes, particularly in healthcare and business, by improving decision accuracy. Quick data processing and operational improvements demonstrate GenAI’s efficiency, particularly in judicial adjudication and teaching. GenAI customization provides personalized solutions, especially in healthcare and tourism; however, strong ethical measures must be implemented to ensure fairness and minimize bias.

Disclosure

This manuscript has been edited with the assistance of GPT-4o for grammar correction and stylistic improvements. All content, intellectual contributions, and final interpretations remain the responsibility of the author, who has thoroughly reviewed and approved the final version of the manuscript.

CRediT authorship contribution statement

Mousa Albashrawi: Writing – review & editing, Writing – original draft, Validation, Supervision, Project administration, Methodology.

References
[Abdelwahed, 2024]
N.A.A. Abdelwahed.
Recognizing the role of ChatGPT in decision-making and recognition of mental health disorders among entrepreneurs.
[Ahmad et al., 2024]
K. Ahmad, A. Rozhok, R. Revetria.
Supply chain resilience in SMEs: Integration of generative AI in decision-making framework.
Proceedings of the 2024 international conference on machine intelligence and smart innovation (ICMISI 2024), pp. 295-299 http://dx.doi.org/10.1109/ICMISI61517.2024.10580495
[Al-Busaidi et al., 2024]
A.S. Al-Busaidi, R. Raman, L. Hughes, M.A. Albashrawi, T. Malik, Y.K. Dwivedi, P. Walton.
Redefining boundaries in innovation and knowledge domains: Investigating the impact of generative artificial intelligence on copyright and intellectual property rights.
Journal of Innovation & Knowledge, 9 (2024),
[Beilby and Hammarberg, 2024]
K. Beilby, K. Hammarberg.
ChatGPT: A reliable fertility decision-making tool?.
Human Reproduction, 39 (2024), pp. 443-447
[Borna et al., 2024]
S. Borna, C.A. Gomez-Cabello, S.M. Pressman, S.A. Haider, A.J. Forte.
Comparative analysis of large language models in emergency plastic surgery decision-making: The role of physical exam data.
Journal of Personalized Medicine, 14 (2024),
[Brügge et al., 2024]
E. Brügge, S. Ricchizzi, M. Arenbeck, M.N. Keller, L. Schur, W. Stummer, D. Darici.
Large language models improve clinical decision-making of medical students through patient simulation and structured feedback: A randomized controlled trial.
BMC Medical Education, 24 (2024),
[Bukar et al., 2024a]
U.A. Bukar, M.S. Sayeed, S. Fatimah Abdul Razak, S. Yogarayan, R. Sneesl.
Decision-making framework for the utilization of generative artificial intelligence in education: A case study of ChatGPT.
IEEE Access, 12 (2024), pp. 95368-95389
[Bukar et al., 2024b]
U.A. Bukar, M.S. Sayeed, S.F.A. Razak, S. Yogarayan, O.A. Amodu.
An integrative decision-making framework to guide policies on regulating ChatGPT usage.
PeerJ Computer Science, 10 (2024),
[Camacho, 2024]
S. Camacho.
A needed (r)evolution to improve colorectal cancer screening: Binding patient-centered/shared decision-making medicine with large language models in an emerging knowledge society.
[Cardoso et al., 2024]
A.G. Cardoso, E. Chan, L. Quintão, C. Pereira.
Generative artificial intelligence and legal decision-making.
Global Trade and Customs Journal, 19 (2024), pp. 710-730
[Carullo, 2023]
G. Carullo.
Large language models for transparent and intelligible AI-assisted public decision-making.
[Chen et al., 2024a]
M. Chen, Z. Tao, W. Tang, T. Qin, R. Yang, C. Zhu.
Enhancing emergency decision-making with knowledge graphs and large language models.
International Journal of Disaster Risk Reduction, (2024), pp. 113
[Chen et al., 2024b]
R. Chen, D. Zeng, Y. Li, R. Huang, D. Sun, T. Li.
Evaluating the performance and clinical decision-making impact of ChatGPT-4 in reproductive medicine.
International Journal of Gynecology and Obstetrics, (2024),
[Christensen et al., 2024]
J. Christensen, J.M. Hansen, P. Wilson.
Understanding the role and impact of generative artificial intelligence (AI) hallucination within consumers’ tourism decision-making processes.
Current Issues in Tourism, (2024),
[Civettini et al., 2024]
I. Civettini, A. Zappaterra, B.M. Granelli, G. Rindone, A. Aroldi, S. Bonfanti, F. Colombo, M. Fedele, G. Grillo, M. Parma, P. Perfetti, E. Terruzzi, C. Gambacorti-Passerini, D. Ramazzotti, F. Cavalca.
Evaluating the performance of large language models in haematopoietic stem cell transplantation decision-making.
British Journal of Haematology, 204 (2024), pp. 1523-1528
[Ding et al., 2024]
P. Ding, J. Liu, M. Sun, L. Li, H. Liu.
Enhancing computational processing performance for generative AI large models with autonomous decision-making in Metaverse applications.
Proceedings of the 2024 IEEE international conference on metaverse computing, networking, and applications (MetaCom 2024) (pp. 253–258), http://dx.doi.org/10.1109/MetaCom62920.2024.00048
[Dwivedi, 2025]
Y.K. Dwivedi.
Generative Artificial intelligence (GenAI) in entrepreneurial education and practice: Emerging insights, the GAIN Framework, and research agenda.
International Entrepreneurship and Management Journal, 21 (2025), pp. 1-21
[Dwivedi et al., 2024]
Y.K. Dwivedi, T. Malik, L. Hughes, M.A. Albashrawi.
Scholarly discourse on GenAI’s impact on academic publishing.
Journal of Computer Information Systems, (2024), pp. 1-16
[Ebrahimian et al., 2023]
M. Ebrahimian, B. Behnam, N. Ghayebi, E. Sobhrakhshankhah.
ChatGPT in Iranian medical licensing examination: Evaluating the diagnostic accuracy and decision-making capabilities of an AI-based model.
BMJ Health and Care Informatics, 30 (2023),
[Garg et al., 2024]
S. Garg, A. Ahmad, D.O. Madsen.
Academic writing in the age of AI: Comparing the reliability of ChatGPT and Bard with Scopus and Web of Science.
Journal of Innovation & Knowledge, 9 (2024),
[Gholami, 2024]
H. Gholami.
Artificial intelligence techniques for sustainable reconfigurable manufacturing systems: An AI-powered decision-making application using large language models.
Big Data and Cognitive Computing, 8 (2024),
[Gloria et al., 2024]
B. Gloria, J. Melsbach, S. Bienert, D. Schoder.
Real-GPT: Efficiently tailoring LLMs for informed decision-making in the real estate industry.
Journal of Real Estate Portfolio Management, (2024),
[Goktas, 2024]
P. Goktas.
Ethics, transparency, and explainability in generative AI decision-making systems: A comprehensive bibliometric study.
Journal of Decision Systems, (2024),
[Griewing et al., 2023]
S. Griewing, N. Gremke, U. Wagner, M. Lingenfelder, S. Kuhn, J. Boekhoff.
Challenging ChatGPT 3.5 in senology—An assessment of concordance with breast cancer tumor board decision-making.
Journal of Personalized Medicine, 13 (2023),
[Griewing et al., 2024]
S. Griewing, J. Knitza, J. Boekhoff, C. Hillen, F. Lechner, U. Wagner, M. Wallwiener, S. Kuhn.
Evolution of publicly available large language models for complex decision-making in breast cancer care.
Archives of Gynecology and Obstetrics, 310 (2024), pp. 537-550
[Gumilar et al., 2024]
K.E. Gumilar, B.R. Indraprasta, A.S. Faridzi, B.M. Wibowo, A. Herlambang, E. Rahestyningtyas, M. Tan.
Assessment of large language models (LLMs) in decision-making support for gynecologic oncology.
Computational and Structural Biotechnology Journal, 23 (2024), pp. 4019-4026
[Haemmerli et al., 2023]
J. Haemmerli, L. Sveikata, A. Nouri, A. May, K. Egervari, C. Freyschlag, J.A. Lobrinus, D. Migliorini, S. Momjian, N. Sanda, K. Schaller, S. Tran, J. Yeung, P. Bijlenga.
ChatGPT in glioma adjuvant therapy decision-making: Ready to assume the role of a doctor in the tumor board?.
BMJ Health and Care Informatics, 30 (2023),
[Hager et al., 2024]
P. Hager, F. Jungmann, R. Holland, K. Bhagat, I. Hubrecht, M. Knauer, J. Vielhauer, M. Makowski, R. Braren, G. Kaissis, D. Rueckert.
Evaluation and mitigation of the limitations of large language models in clinical decision-making.
Nature Medicine, 30 (2024), pp. 2613-2622
[Hao et al., 2024]
X. Hao, E. Demir, D. Eyers.
Exploring collaborative decision-making: A quasi-experimental study of human and generative AI interaction.
Technology in Society, 78 (2024),
[Hiranchan and Kiattisin, 2024]
S. Hiranchan, S. Kiattisin.
The current state of generative AI prompt framework design for enhancing utility in organizational decision-making.
Proceedings of the 5th technology innovation management and engineering science international conference (TIMES-iCON 2024), http://dx.doi.org/10.1109/TIMES-ICON61890.2024.10630713
[Huang et al., 2023]
Y. Huang, A. Gomaa, S. Semrau, M. Haderlein, S. Lettmaier, T. Weissmann, J. Grigo, H.B. Tkhayat, B. Frey, U. Gaipl, L. Distel, A. Maier, R. Fietkau, C. Bert, F. Putz.
Benchmarking ChatGPT-4 on a radiation oncology in-training exam and Red journal gray zone cases: Potentials and challenges for AI-assisted medical education and decision-making in radiation oncology.
Frontiers in Oncology, 13 (2023),
[Huo et al., 2024]
B. Huo, E. Calabrese, P. Sylla, S. Kumar, R.C. Ignacio, R. Oviedo, I. Hassan, B.J. Slater, A. Kaiser, D.S. Walsh, W. Vosburg.
The performance of artificial intelligence large language model-linked chatbots in surgical decision-making for gastroesophageal reflux disease.
Surgical Endoscopy, 38 (2024), pp. 2320-2330
[Jiang et al., 2024]
Y. Jiang, Z. Li, C.L. Philip Chen.
Research on financial big data collection and intelligent decision-making system based on multimodal large language model.
Proceedings of the 2024 international conference on fuzzy theory and its applications (iFUZZY 2024), http://dx.doi.org/10.1109/iFUZZY63051.2024.10662884
[Joshi et al., 2024]
S.G. Joshi, C.H. Tolloczko, S. Wenaas, O.J. Holm, M.L.F. Langved.
Investigating how generative AI affects decision-making in participatory design.
ACM international conference proceeding series, http://dx.doi.org/10.1145/3679318.3685384
[Ke et al., 2024]
Y. Ke, R. Yang, S.A. Lie, T.X.Y. Lim, Y. Ning, I. Li, H.R. Abdullah, D.S.W. Ting, N. Liu.
Mitigating cognitive biases in clinical decision-making through multi-agent conversations using large language models: Simulation study.
Journal of Medical Internet Research, (2024), pp. 26
[Kim et al., 2024]
J.H. Kim, J. Kim, S. Kim, T.B. Hailu.
Effects of AI ChatGPT on travelers’ travel decision-making.
Tourism Review, 79 (2024), pp. 1038-1057
[Kim et al., 2025]
J.H. Kim, J. Kim, J. Park, C. Kim, J. Jhang, B. King.
When ChatGPT gives incorrect answers: The impact of inaccurate information by generative AI on tourism decision-making.
Journal of Travel Research, 64 (2025), pp. 51-73
[Kromidha and Davison, 2024]
E. Kromidha, R.M. Davison.
Generative AI-augmented decision-making for business information systems.
IFIP advances in information and communication technology (Vol. 719, pp. 46–55), Springer, (2024), http://dx.doi.org/10.1007/978-3-031-67535-5_5
[Kuang et al., 2024]
S. Kuang, Y. Liu, X. Wang, X. Wu, Y. Wei.
Harnessing multimodal large language models for traffic knowledge graph generation and decision-making.
Communications in Transportation Research, 4 (2024),
[Lakkaraju et al., 2023]
K. Lakkaraju, S.E. Jones, S.K.R. Vuruma, V. Pallagani, B.C. Muppasani, B. Srivastava.
LLMs for financial advisement: A fairness and efficacy study in personal decision-making.
Proceedings of the 4th ACM international conference on AI in finance (ICAIF 2023), pp. 100-107 http://dx.doi.org/10.1145/3604237.3626867
[Lawson McLean et al., 2024]
A. Lawson McLean, Y. Wu, A.C. Lawson McLean, V. Hristidis.
Large language models as decision aids in neuro-oncology: A review of shared decision-making applications.
Journal of Cancer Research and Clinical Oncology, 150 (2024), pp. 139
[Lee et al., 2023]
P.C. Lee, S.K. Sharma, S. Motaganahalli, A. Huang.
Evaluating the clinical decision-making ability of large language models using MKSAP-19 cardiology questions.
[Levartovsky et al., 2023]
A. Levartovsky, S. Ben-Horin, U. Kopylov, E. Klang, Y. Barash.
Towards AI-augmented clinical decision-making: An examination of ChatGPT's utility in acute ulcerative colitis presentations.
American Journal of Gastroenterology, 118 (2023), pp. 2283-2289
[Liao et al., 2023]
J. Liao, H. Zhang, H. Qian, Q. Meng, Y. Sun, Y. Sun, W. Song, S. Zhu, J. Gu.
Decision-making in robotic grasping with large language models.
Lecture notes in computer science (Vol. 14271, pp. 424–433), Springer, (2023), http://dx.doi.org/10.1007/978-981-99-6495-6_36
[Lin and Clark, 2024]
Y. Lin, K.D. Clark.
Ethical decision-making, plagiarism, and using ChatGPT in the first-year public speaking course.
Communication Quarterly, 72 (2024), pp. 598-619
[Lin et al., 2024]
J.C. Lin, S.S. Kurapati, D.N. Younessi, I.U. Scott, D.A. Gong.
Ethical and professional decision-making capabilities of artificial intelligence chatbots: Evaluating ChatGPT’s professional competencies in medicine.
Medical Science Educator, 34 (2024), pp. 331-333
[Loya et al., 2023]
M. Loya, D.A. Sinha, R. Futrell.
Exploring the sensitivity of LLMs' decision-making capabilities: Insights from prompt variation and hyperparameters.
Findings of the association for computational linguistics: EMNLP 2023, pp. 3711-3716
[Lukac et al., 2023]
S. Lukac, D. Dayan, V. Fink, E. Leinert, A. Hartkopf, K. Veselinovic, W. Janni, B. Rack, K. Pfister, B. Heitmeir, F. Ebner.
Evaluating ChatGPT as an adjunct for the multidisciplinary tumor board decision-making in primary breast cancer cases.
Archives of Gynecology and Obstetrics, 308 (2023), pp. 1831-1844
[Malloy and Gonzalez, 2024]
T. Malloy, C. Gonzalez.
Applying generative artificial intelligence to cognitive models of decision-making.
Frontiers in Psychology, 15 (2024),
[Marchi et al., 2024]
F. Marchi, E. Bellini, A. Iandelli, C. Sampieri, G. Peretti.
Exploring the landscape of AI-assisted decision-making in head and neck cancer treatment: A comparative analysis of NCCN guidelines and ChatGPT responses.
European Archives of Oto-Rhino-Laryngology, 281 (2024), pp. 2123-2136
[Markova et al., 2024]
O. Markova, I. Muzyka, D. Kuznetsov, Y. Kumchenko, A. Senko.
Enhancing IoT and cyber-physical systems in industry 4.0 through on-premise large language models: Real-time data processing, predictive maintenance, and autonomous decision-making.
CEUR Workshop Proceedings, 3723 (2024), pp. 182-197
[McDarby et al., 2024]
M. McDarby, E.L. Mroz, J. Hahne, C.D. Malling, B.D. Carpenter, P.A. Parker.
Hospice care could be a compassionate choice”: ChatGPT responses to questions about decision making in advanced cancer.
Journal of Palliative Medicine, (2024),
[Miao et al., 2024]
J. Miao, C. Thongprayoon, T. Fülöp, W. Cheungpasitporn.
Enhancing clinical decision-making: Optimizing ChatGPT's performance in hypertension care.
Journal of Clinical Hypertension, 26 (2024), pp. 588-593
[Moravec et al., 2024]
V. Moravec, N. Hynek, B. Gavurova, M. Rigelsky.
Who uses it and for what purpose? The role of digital literacy in ChatGPT adoption and utilisation.
Journal of Innovation & Knowledge, 9 (2024),
[Mykhalko et al., 2023]
Y. Mykhalko, P. Kish, Y. Rubtsova, O. Kutsyn, V. Koval.
From text to diagnose: ChatGPT's efficacy in medical decision-making.
Wiadomości Lekarskie, 76 (2023), pp. 2345-2350
[Nechay et al., 2024]
T.V. Nechay, A.V. Sazhin, K.M. Loban, A.K. Bogomolova, V.V. Suglob, T.R. Beniia.
Efficacy and safety of artificial intelligence-based large language models for decision-making support in herniology: Evaluation by experts and general surgeons.
Pirogov Russian Journal of Surgery, (2024), pp. 6-14
[Palenzuela et al., 2024]
D.L. Palenzuela, J.T. Mullen, R. Phitayakorn.
AI versus MD: Evaluating the surgical decision-making accuracy of ChatGPT-4.
Surgery, 176 (2024), pp. 241-245
[Perona and de la Rosa, 2024]
R. Perona, Y.C. de la Rosa.
Unveiling AI in the courtroom: Exploring ChatGPT’s impact on judicial decision-making through a pilot Colombian case study.
[Pitkäranta and Pitkäranta, 2024]
T. Pitkäranta, L. Pitkäranta.
Bridging human and AI decision-making with LLMs: The RAGADA approach.
Proceedings of the international conference on enterprise information systems (ICEIS 2024), pp. 812-819 http://dx.doi.org/10.5220/0012705000003690
[Rajjoub et al., 2024]
R. Rajjoub, J.S. Arroyave, B. Zaidat, W. Ahmed, M.R. Mejia, J. Tang, J.S. Kim, S.K. Cho.
ChatGPT and its role in the decision-making for the diagnosis and treatment of lumbar spinal stenosis: A comparative analysis and narrative review.
Global Spine Journal, 14 (2024), pp. 998-1017
[Saban and Dubovi, 2024]
M. Saban, I. Dubovi.
A comparative vignette study: Evaluating the potential role of a generative AI model in enhancing clinical decision-making in nursing.
Journal of Advanced Nursing, (2024),
[Salihu et al., 2023]
A. Salihu, M.A. Gadiri, I. Skalidis, D. Meier, D. Auberson, A. Fournier, R. Fournier, D. Thanou, E. Abbé, O. Muller, S. Fournier.
Towards AI-assisted cardiology: A reflection on the performance and limitations of using large language models in clinical decision-making.
EuroIntervention, 19 (2023), pp. 798-801
[Sarangi et al., 2024]
P.K. Sarangi, S. Datta, M.S. Swarup, S. Panda, D.S. Kumar Nayak, A. Malik, A. Datta, H. Mondal.
Radiologic decision-making for imaging in pulmonary embolism: Accuracy and reliability of large language models—Bing, Claude, ChatGPT, and perplexity.
Indian Journal of Radiology and Imaging, 34 (2024), pp. 653-660
[Sblendorio et al., 2024]
E. Sblendorio, V. Dentamaro, A. Lo Cascio, F. Germini, M. Piredda, G. Cicolini.
Integrating human expertise & automated methods for a dynamic and multi-parametric evaluation of large language models’ feasibility in clinical decision-making.
International Journal of Medical Informatics, (2024), pp. 188
[Scaioli et al., 2024]
G. Scaioli, G. Lo Moro, F. Conrado, L. Rosset, F. Bert, R. Siliquini.
Exploring the potential of ChatGPT for clinical reasoning and decision-making: A cross-sectional study on the Italian Medical Residency Exam.
Annali dell'Istituto Superiore di Sanita, 59 (2024), pp. 267-270
[Schmidl et al., 2024]
B. Schmidl, T. Hütten, S. Pigorsch, F. Stögbauer, C.C. Hoch, T. Hussain, B. Wollenberg, M. Wirth.
Assessing the role of advanced artificial intelligence as a tool in multidisciplinary tumor board decision-making for recurrent/metastatic head and neck cancer cases – the first study on ChatGPT 4.0 and a comparison to ChatGPT 4.0.
Frontiers in Oncology, 14 (2024),
[Schmidt et al., 2024]
H.G. Schmidt, J.I. Rotgans, S. Mamede.
Bias sensitivity in diagnostic decision-making: Comparing ChatGPT with residents.
Journal of General Internal Medicine, (2024),
[Schumacher et al., 2024]
I. Schumacher, V.M.M. Bühler, D. Jaggi, J. Roth.
Artificial intelligence-derived large language model in decision-making process in uveitis.
International Journal of Retina and Vitreous, 10 (2024),
[Seita and Kurahashi, 2024]
H. Seita, S. Kurahashi.
Research on improving decision-making efficiency with ChatGPT.
[Sel et al., 2024]
B. Sel, P. Shanmugasundaram, M. Kachuee, K. Zhou, R. Jia, M. Jin.
Skin-in-the-game: Decision making via multi-stakeholder alignment in LLMs.
Proceedings of the annual meeting of the association for computational linguistics (Vol. 1, pp. 13921–13959),
[Shan and Michel, 2024]
T. Shan, K. Michel.
Generative AI with GOAP for fast-paced dynamic decision-making in game environments.
Proceedings of the IEEE conference on computational intelligence and games (CIG 2024), http://dx.doi.org/10.1109/CoG60054.2024.10645549
[Shi et al., 2023]
W. Shi, Y. Zhuang, Y. Zhu, H. Iwinski, M. Wattenbarger, M.D. Wang.
Retrieval-augmented large language models for adolescent idiopathic scoliosis patients in shared decision-making.
Proceedings of the 14th ACM conference on bioinformatics, computational biology, and health informatics (ACM-BCB 2023), http://dx.doi.org/10.1145/3584371.3612956
[Siddik et al., 2024]
A.B. Siddik, Y. Li, A.M. Du.
Unlocking funding success for generative AI startups: The crucial role of investor influence.
Finance Research Letters, (2024), pp. 69
[Skryd and Lawrence, 2024]
A. Skryd, K. Lawrence.
ChatGPT as a tool for medical education and clinical decision-making on the wards: Case study.
JMIR Formative Research, 8 (2024),
[Socol De La Osa and Remolina, 2024]
D.U. Socol De La Osa, N. Remolina.
Artificial intelligence at the bench: Legal and ethical challenges of informing - or misinforming - judicial decision-making through generative AI.
Data and Policy, 6 (2024), pp. e59
[Solaiman, 2024]
B. Solaiman.
Generative artificial intelligence (GenAI) and decision-making: Legal & ethical hurdles for implementation in mental health.
International Journal of Law and Psychiatry, (2024), pp. 97
[Stergiou and Nella, 2024]
D. Stergiou, A. Nella.
ChatGPT and tourist decision-making: An accessibility–diagnosticity theory perspective.
International Journal of Tourism Research, 26 (2024),
[Teixeira-Marques et al., 2024]
F. Teixeira-Marques, N. Medeiros, F. Nazaré, S. Alves, N. Lima, L. Ribeiro, R. Gama, P. Oliveira.
Exploring the role of ChatGPT in clinical decision-making in otorhinolaryngology: A ChatGPT designed study.
European Archives of Oto-Rhino-Laryngology, 281 (2024), pp. 2023-2030
[Trillo et al., 2024a]
J.R. Trillo, F.J. Cabrerizo, I.J. Perez, J.A. Morente-Molinera, E. Herrera-Viedma.
A new consensus reaching method for group decision-making based on the large language model Gemini for detecting hostility during the discussion process.
Proceedings of the IEEE conference on evolving and adaptive intelligent systems (EAIS 2024), http://dx.doi.org/10.1109/EAIS58494.2024.10570029
[Trillo et al., 2024b]
J.R. Trillo, M.Á. Martínez, S. Zadrozny, J. Kacprzyk, E. Herrera-Viedma, F.J. Cabrerizo.
A group decision-making approach leveraging preference relations derived from large language models.
Procedia Computer Science, 242 (2024), pp. 452-459
[Ulloa et al., 2023]
J.V. Ulloa, S.K. Valenzuela, C.R. Altamirano, G.U. Schmied.
Artificial intelligence-based decision-making: Can ChatGPT replace a multidisciplinary tumour board?.
British Journal of Surgery, 110 (2023), pp. 1543-1544
[Wong et al., 2023]
I.A. Wong, Q.L. Lian, D. Sun.
Autonomous travel decision-making: An early glimpse into ChatGPT and generative AI.
Journal of Hospitality and Tourism Management, 56 (2023), pp. 253-263
[Wu et al., 2024]
T. Wu, J. Li, J. Bao, Q. Liu.
ProcessCarbonAgent: A large language models-empowered autonomous agent for decision-making in manufacturing carbon emission management.
Journal of Manufacturing Systems, 76 (2024), pp. 429-442
[Xiang et al., 2024]
X. Xiang, J. Xue, L. Zhao, Y. Lei, C. Yue, K. Lu.
Real-time integration of fine-tuned large language model for improved decision-making in reinforcement learning.
Proceedings of the international joint conference on neural networks (IJCNN 2024), http://dx.doi.org/10.1109/IJCNN60899.2024.10650538
[Xu et al., 2024]
W. Xu, M.C. Kotecha, D.A. McAdams.
How good is ChatGPT? An exploratory study on ChatGPT’s performance in engineering design tasks and subjective decision-making.
Proceedings of the design society, pp. 2307-2316 http://dx.doi.org/10.1017/pds.2024.233
[Yadav et al., 2024]
D. Yadav, H. Para, K. Sandhu, P. Selvakumar.
Enhancing response generation systems: Knowledge graph & generative AI synergy for business communication and strategic decision-making.
Proceedings of the 2024 international conference on electrical, computer, and energy technologies (ICECET 2024), http://dx.doi.org/10.1109/ICECET61485.2024.10698632
[Zhang et al., 2023]
H. Zhang, J. Li, Y. Wang, Y. Songi.
Integrating automated knowledge extraction with large language models for explainable medical decision-making.
Proceedings of the 2023 IEEE international conference on bioinformatics and biomedicine, pp. 1710-1717 http://dx.doi.org/10.1109/BIBM58861.2023.10385557
[Zhang and Liang, 2024]
Q. Zhang, Y. Liang.
Comments on “ChatGPT and its role in the decision-making for the diagnosis and treatment of lumbar spinal stenosis: A comparative analysis and narrative review.
Global Spine Journal, 14 (2024), pp. 1452-1453
[Zhong et al., 2023]
Y. Zhong, K. Lan, S. Fong, D. Wong.
Intelligent assisted decision-making framework for domain-specific advice using large language models.
Proceedings of the 2023 international conference on digital data processing (DDP 2023), pp. 24-30 http://dx.doi.org/10.1109/DDP60485.2023.00015
Download PDF
Article options