ѻý

AI in Healthcare: Navigating the Crossroads of Innovation and Equity

<ѻý class="mpt-content-deck">— Tech solutions should be aimed at boosting efficiency and care access
MedpageToday
A computer rendering of robot hands touching data on a holographic display.

highlighted the escalating concerns with the adoption of new artificial intelligence (AI) tools aimed at streamlining operations for tasks traditionally seen as expensive and complex, like recruitment processes. Critics argue that the exact nature and extent of potential harms remain elusive. Yet, when it comes to the healthcare industry, the ramifications are both and .

Our ill patients and aging population demand a more efficient and effective healthcare system. In response, administrators and policymakers increasingly look to AI for solutions, such as automated diagnostics, predictive analytics, provider scheduling, and patient management systems. However, deploying these technologies prompts vital questions about their impact on healthcare delivery, especially in terms of provider burnout and ensuring to quality care.

These concerns are a constant reality. For example, if an AI system designed to facilitate patient care encounters errors in processing language differences, this miscommunication between a patient with limited English proficiency, the pharmacy, and the healthcare system could delay a critical medication refill, leading to a potential AI-induced drug safety event. This type of incident highlights the systemic challenges within healthcare technology environments, where accountability often shifts between digital and human systems, leaving essential systemic issues unaddressed.

This example serves as a snapshot of broader issues within healthcare technology, which often has greater consequences for minority groups, and underscores the critical need for more inclusive and equitable healthcare technologies. Patients from minoritized communities may experience significant delays in accessing healthcare services due to these technological disparities. Regrettably, these crucial data insights might be overlooked by administrators who may need to pressure healthcare providers to manage an increasing patient load, inadvertently neglecting the complexity of care required.

In this context, the potential of AI in healthcare is tremendous, offering the possibility to predict patient complexity, enhance care appropriateness, streamline operations, boost patient satisfaction and outcomes, and level the playing field. However, without a thoughtful implementation strategy, AI risks .

With over a decade in , my career places me at the heart of healthcare's dynamic field. Despite optimism for technological progress, past experiences delivering care in underserved regions haunt me. Arriving in the U.S. over a decade ago, hopeful for global healthcare standards, I've seen our nation battling increasing rates of provider burnout while dealing with nursing staff shortages and the challenge of managing persistent disparities in access to care.

This issue is underscored by my observations when interacting with healthcare executives from across the U.S. at professional conferences, where the emphasis on AI sometimes seems more driven by revenue enhancement than by the desire to improve patient outcomes, teamwork, or healthcare experiences. For instance, when choosing between investing in a software that could alleviate the of documentation for healthcare providers (a major contributor to burnout) and a software aimed at cutting nursing costs by restricting work hours (thus avoiding after-hours payment), the latter was chosen by some executives, with little regard for its potential impact on patient care. I fear that this could be implemented in multiple institutions across the U.S., as we are already facing nursing workforce shortages, taking a toll on our mental health.

The success of AI in healthcare must be measured not only by its financial impact but also by its contribution to patient outcomes and healthcare equity. This approach necessitates the development of AI algorithms using , comprehensive data sets and ongoing assessments to ensure their effectiveness in diverse real-world settings.

While the argument exists that revenue from AI-driven healthcare systems could fund separate access and equity projects, this perspective is fraught with risk. Prioritizing financial gains initially may reinforce biases or neglect vulnerable populations, a concern supported by . Such damage might not be offset by subsequent equity investments. Additionally, reports have linked inefficiencies in healthcare systems to the of healthcare providers, emphasizing the need for solutions that prioritize the well-being of both patients and providers.

My vision for AI in healthcare centers on fostering equity and provider well-being through value-based care models. This entails prioritizing patient outcomes and provider satisfaction, guaranteeing quality care for all, irrespective of background. To actualize this, we must choose AI solutions that meet , including unbiased data use. Implementation should drive outcome improvements and cater to diverse providers and patients. Continuous monitoring is crucial to identify and address , particularly those impacting marginalized groups.

The human element in healthcare is crucial amid skyrocketing provider burnout rates (an all-time high of ). AI should be a tool to cut inefficiencies, allowing more time for human connections and team building, not perpetuating blame. AI implementation must consider the impact on provider workflows, workloads, and morale, aiming to measure its potential to reduce burnout, enhance job satisfaction, and improve access to care.

Healthcare stakeholders, including administrators, providers, patients, sponsors, must adopt AI with a clear vision and a commitment to boost efficiency and equity. There needs to be a balanced approach that aligns technological advancements with understanding their impacts, especially on vulnerable patients, ensuring no patient is left behind. Providers should not feel overwhelmed, understanding that AI tools are meant to assist, not constrain.

is a clinical neurologist at Massachusetts General Hospital and Harvard Medical School in Boston, with expertise in biomedical informatics, advanced general neurology, neurophysiology, epidemiology, and health services research. She is also an OpEd project fellow and director of the Center for Value-based Healthcare and Sciences with Massachusetts General Hospital.