Class-action lawsuit accuses Humana of AI-driven, post-acute care declare denials
A category motion lawsuit towards Humana Inc. was filed in Kentucky on Tuesday. (NYSE: HUM). The corporate is the most recent insurer to return below hearth for its alleged use of AI to dictate or deny post-acute care.
Beneficiaries of Humana Medicare Benefit (MA) particularly filed the lawsuit, alleging that the corporate illegally used algorithms to restrict post-acute care companies. The algorithm device in query is named 'nH Predict'.
Humana, based mostly in Louisville, is likely one of the nation's largest insurers. It has nearly 6 million MA beneficiaries. The corporate additionally owns CenterWell House Well being, one of many nation's largest residence well being care suppliers.
“Humana systematically deploys the AI algorithm to prematurely and in dangerous religion cease fee for well being care companies for older adults with critical diseases and accidents,” the lawsuit says. “These older sufferers are left with monumental medical debt or with out the medical care they want.”
After STAT Information carried out an investigation into Medicare Benefit insurers and their use of algorithms to disclaim fee for care, a number of class motion lawsuits had been filed.
Not too long ago, a lawsuit was filed towards UnitedHealth Group over the denial actions of its organizer – naviHealth – in Minnesota. Cigna was additionally hit with a category motion lawsuit in Connecticut over its declare denials via the PxDx device.
House care sufferers are affected by these denials, as are healthcare suppliers. Suppliers have lengthy been involved about “middlemen” influencing care and fee for that care.
“We’ve got all of the capabilities that organizers have,” Chris Gerard, the previous CEO of Amedisys Inc., advised me. (Nasdaq: AMED), in 2022 to House Well being Care Information. “The one factor they’ll do for plans that we're not going to do is handle a community of suppliers. We will completely do what they do with regards to utilization administration. We will obtain the identical outcomes. We will generate financial savings for the plan. [Given that]it makes completely no sense for us to cooperate [them].”
The lawsuits additionally shine a highlight on the usage of AI in healthcare on the whole. AI is already practically ubiquitous in healthcare, however its novelty raises moral questions.
The denial of claims by the aforementioned AI instruments has reportedly pressured sufferers to pay 1000’s of {dollars} out of pocket for care in some circumstances.
Humana advised House Well being Care Information that it “doesn’t touch upon pending litigation,” however did elaborate on its use of “augmented intelligence.”
“I can verify that at Humana we use quite a lot of instruments, together with augmented intelligence, to expedite and approve utilization administration requests and guarantee sufferers obtain high-quality, secure and environment friendly care,” a Humana spokesperson stated. “By definition, augmented intelligence permits for 'human within the loop' resolution making when AI is used. Protection choices are made based mostly on sufferers' well being care wants, the medical judgment of physicians and physicians, and pointers established by CMS. You will need to be aware that hostile protection choices are made solely by doctor medical administrators.”