We Must Not Let AI ‘Pull the Doctor Out of the Visit’ for Low-Income Patients | Leah Goodridge and Oni Blackstock

We Must Not Let AI ‘Pull the Doctor Out of the Visit’ for Low-Income Patients | Leah Goodridge and Oni Blackstock

The Guardian AI
The Guardian AIJan 25, 2026

Why It Matters

AI‑enabled diagnostics risk widening health disparities for low‑income patients, while unchecked deployment threatens patient autonomy and legal accountability.

Key Takeaways

  • Akido Labs uses AI to replace doctors in visits
  • AI tools misdiagnose Black, Latinx, Medicaid patients
  • Patients often unaware AI assists their care
  • AI‑driven insurance decisions spark lawsuits
  • Bias‑laden data amplifies existing health inequities

Pulse Analysis

The rapid integration of generative AI into clinical workflows is driven by systemic pressures—overcrowded hospitals, clinician burnout, and profit‑centric health systems. Startups like Akido Labs promise efficiency by letting medical assistants record encounters while an algorithm suggests diagnoses, ostensibly freeing physicians for higher‑order tasks. For low‑income and unhoused populations, however, this model substitutes a critical human touch with opaque software, raising concerns about quality of care and the erosion of trust in already fragile provider‑patient relationships.

Research consistently shows that AI models inherit and magnify biases present in their training data. A 2021 Nature Medicine study revealed systematic under‑diagnosis of Black, Latinx, female, and Medicaid patients in chest‑X‑ray algorithms, while a 2024 breast‑cancer screening analysis found higher false‑positive rates for Black women. These inaccuracies are not merely statistical quirks; they translate into delayed treatment, unnecessary procedures, and reinforced health inequities. Moreover, many patients are not informed that AI is listening and influencing clinical decisions, echoing historic abuses of medical experimentation without consent.

Legal scrutiny is intensifying as AI’s reach expands beyond diagnosis to coverage determinations. Cases against UnitedHealthcare and Humana illustrate how algorithmic errors can deny life‑saving care, prompting courts to allow claims to proceed. Policymakers must therefore mandate transparent AI use, rigorous bias testing, and community‑led oversight, especially for vulnerable groups. Until robust safeguards are in place, the healthcare industry should prioritize human‑centered care over unproven technological shortcuts, ensuring that AI augments rather than replaces the clinician’s judgment.

We must not let AI ‘pull the doctor out of the visit’ for low-income patients | Leah Goodridge and Oni Blackstock

Comments

Want to join the conversation?

Loading comments...