A trio of health policy analysts from the University of Pennsylvania and the University of California has published a Policy Forum piece in the journal Science calling for more stringent rules for the introduction of AI medical applications. In their paper, Ravi Parikh, Ziad Obermeyer and Amol Navathe outline five standards they believe should be instituted as part of allowing AI medical applications to be used for medical purposes.
The authors note that the use of AI-based tools to diagnose medical conditions or to provide predictions for outcomes based on different treatment approaches is still very new. It has only been in the past few years that AI-based algorithms and tools that use them have been developed. Because of that, rules for their use have not been well established. In their paper, the researchers suggest five standards that should be implemented to safeguard patients who are part of medical treatment that involves the use of AI applications or devices.
The first involves establishing endpoints that are meaningful—benefits should be clearly identifiable and thus subject to validation by the FDA in the same way that drugs and other devices have been for many years. The second standard involves establishing benchmarks that are appropriate to the area in which they are applied, allowing their usefulness and quality to be properly evaluated. The third standard involves ensuring that variable input specifications are clear so that they can be used by more than one institution when testing a new application or device. The fourth standard involves possible interventions associated with findings by AI systems and whether they are appropriate and successful. And finally, the fifth standard would be the implementation of regular rigorous audits—a standard that has been used when introducing new drugs for many years. Auditing AI applications takes on added importance when considering that the data that underlie their predictive abilities would change over time.
Source: Read Full Article