Imagine a future where a computer algorithm decides whether you receive the medical treatment your doctor recommends. This is not science fiction; it’s a reality Medicare is testing right now. In a bold move, Medicare has launched a six-year pilot program that could fundamentally alter how millions of Americans access healthcare. Dubbed the Wasteful and Inappropriate Service Reduction Model (WISER), this initiative aims to curb unnecessary spending by using artificial intelligence (AI) to review and potentially deny certain medical procedures and devices for those enrolled in traditional Medicare. But here’s where it gets controversial: while this could save taxpayers billions, it also raises serious concerns about whether patients might be denied essential care.
Traditional Medicare, a government-run insurance program for Americans over 65 or with disabilities, covers roughly half of the 67 million people insured through Medicare. The other half are enrolled in Medicare Advantage plans, administered by private companies. Starting in January 2026, the WISER pilot will affect traditional Medicare enrollees in six states: Arizona, New Jersey, Ohio, Oklahoma, Texas, and Washington. During this trial, healthcare providers must seek prior authorization for 14 specific procedures and devices, such as steroid injections for pain management and incontinence-control devices. The AI system will flag and deny requests it deems unnecessary or harmful, a practice already common in Medicare Advantage plans.
But this is the part most people miss: Traditional Medicare has historically avoided prior authorization requirements, allowing doctors to recommend treatments without bureaucratic hurdles. This pilot marks a significant shift, introducing an extra layer of scrutiny that could delay or even block necessary care. While the goal of reducing wasteful spending is commendable, the risk of denying essential treatments is a serious concern. Health economists like us, who have studied both Medicare and AI in healthcare, believe this pilot could indeed save money, but it must be closely monitored to ensure patient health isn’t compromised.
Here’s the catch: The tech companies hired to manage these AI-driven decisions are paid a percentage of what Medicare would have spent on denied treatments. This creates a financial incentive to deny more requests, potentially prioritizing profits over patient care. While Medicare will monitor for inappropriate denials, the system’s accuracy and fairness remain unproven. Past research shows that prior authorization often leads to fewer services for patients, and early evidence suggests AI-aided decisions result in higher denial rates and reduced healthcare use.
So, what’s at stake? If successful, this pilot could expand to more states and procedures, reshaping Medicare’s future. But the trade-offs are significant. Tech companies stand to gain financially, while healthcare providers face increased paperwork and reduced payments. Patients could benefit from avoiding unnecessary treatments, but they might also lose access to essential care if the AI fails to distinguish between the two. Taxpayers could see long-term savings, but at what cost to the quality of care?
Here’s a thought-provoking question for you: Is it ethical to entrust life-altering healthcare decisions to algorithms, especially when financial incentives might skew their outcomes? Share your thoughts in the comments—we’d love to hear your perspective on this complex and controversial issue.