Healthcare AI takes another jab to the jaw

Last week a literature review showed none of 62 high-quality medical AI models ready for translation from academic research to clinical practice. Now comes a similar but separate study confirming the depth of the dashed hopes.

Reporting their findings in Science Translational Medicine, the researchers behind the second exercise found just 23% of healthcare machine-learning studies were reproducible with differing datasets.

By comparison, 80% of computer vision studies and 58% of NLP studies had such conceptual reproducibility.

Equally confounding, 55% of machine learning in healthcare papers used public datasets and made their code available. Computer vision and NLP each clocked in at close to 90% on those scores.

IEEE Spectrum takes a quick look at both literature reviews side by side.

“Healthcare is an especially challenging area for machine learning research because many datasets are restricted due to health privacy concerns and even experts may disagree on a diagnosis for a scan or patient,” writes freelance journalist Megan Scudellari. “Still, researchers are optimistic that the field can do better.”

Read the whole thing.

Dave Pearson

Dave P. has worked in journalism, marketing and public relations for more than 30 years, frequently concentrating on hospitals, healthcare technology and Catholic communications. He has also specialized in fundraising communications, ghostwriting for CEOs of local, national and global charities, nonprofits and foundations.

Around the web

Compensation for heart specialists continues to climb. What does this say about cardiology as a whole? Could private equity's rising influence bring about change? We spoke to MedAxiom CEO Jerry Blackwell, MD, MBA, a veteran cardiologist himself, to learn more.

The American College of Cardiology has shared its perspective on new CMS payment policies, highlighting revenue concerns while providing key details for cardiologists and other cardiology professionals. 

As debate simmers over how best to regulate AI, experts continue to offer guidance on where to start, how to proceed and what to emphasize. A new resource models its recommendations on what its authors call the “SETO Loop.”