Acceso abierto

A Counterblast to Pessimists and Naysayers – Intelligent Echocardiography Remains the Foundation Stone of Evidence-Based Clinical Cardiology

   | 26 dic 2023

Cite

Combatting misconceptions about artificial intelligence
It is artificial, but it is not intelligent

All diagnostic imaging systems now generate a huge quantity of digital data from each study. Images are processed and post-processed to improve their quality, using software that is specific to each company because it has been developed by its own engineers. Algorithms are protected as intellectual property, so they are not shared with other companies or made available to clinicians.

Nowadays, a diagnostic study in an individual patient usually involves collecting imaging data in three dimensions over time, with the goal not of making simple single measurements but of analyzing many aspects of function, perhaps both at rest and during stress. The challenges of coping with this mass of data are compounded when images are collected and combined, perhaps from many patients. It is entirely understandable that we need new methods to make sense of such vast imaging biobanks.

There is a strong argument that AI consists of advanced statistical methodologies[1] that have had to be developed because previous methods can no longer cope with the avalanche of data. That concept is much more prosaic than the current fashion and hype which brand AI as a paradigm shift in technology. The vocabulary that is now used to describe AI is misleading and anthropomorphic, since the plain truth is that no computer can “understand” what is does.[2] A computer is not a conscious organism but a machine whose sole function is to perform the tasks that it has been programmed to perform. If that task happens to be running an AI algorithm, it still consists only of instructions to process bits (of binary code) and to perform FLOPS (floating point operations per second) according to instructions prepared by a human being. If an algorithm can “self-learn,” it is only because it has been programmed to do that. The real intelligence is always human, not artificial.

Generalizations about AI are misleading

It is misguided to refer to AI as if it is a single phenomenon, since the term encompasses anything from “general” AI (referring to a program intended to mimic human performance by generating “thoughts” and replicating actions) to machine learning (ML) algorithms to deep neural networks. For applications in medicine, the definition of the World Health Organisation may be most useful, namely “the ability of algorithms to learn from data so that they can perform automated tasks.”[3] Note the absence of the word intelligence from this description.

Smart cardiologists should avoid talking about AI in broad terms, without qualification.

Furthermore, using AI is not a primary objective, whereas using a specific type of AI when it is the best method of performing a particular task or answering a particular question is laudable, perhaps even essential. It is wiser to use a precise term for each specific application.

General AI is not a threat to clinical practice

Recent public concern and debate around the world has been driven by the rapid development of “generative” AI tools for language processing (such as the “Chat Generative Pre-trained Transformer,” or ChatGPT, in particular) with their perceived and unknown risks—but as far as we know, they are not used by diagnostic imaging systems. In addition, true general AI does not yet exist, and it seems very unlikely that it would ever be implemented into some aspect of clinical practice because of the implications of ceding human control.

Software including AI algorithms with a medical function, such as making a diagnosis, planning treatment, or predicting outcomes, is designated in the European Union as a medical device (SaMD, or “software as a medical device”).[4] Whether it is a standalone program or it is incorporated within an implantable device, it has to be approved by a notified body and given a CE mark (for Conformité Européenne) before it can be implemented in practice. High-risk SaMD that is used in critical clinical situations, when any erroneous operation could lead to potentially serious consequences for the patient, will only be approved if it has undergone a satisfactory clinical evaluation with evidence of substantial benefit that outweighs any risks. All major jurisdictions for medical devices are developing regulations to ensure that high-risk AI or ML must meet certain minimum standards.[5] The engagement and collaboration of expert clinicians in supporting regulatory decisions can ensure that medical ML algorithms are approved only when they are needed and will benefit patients. The European Society of Cardiology is leading a European consortium which has amongst its tasks a project to define the criteria on which such decisions should be based.[6,7]

Exploiting the clinical potential of AI

The vast majority of AI medical devices which have been approved by regulators are software applications for diagnostic imaging that are already in routine use for tasks such as optimizing image quality, detecting and labeling structures (called “segmentation” by AI engineers), and performing measurements. Systematic reviews suggest that ML imaging tools can outperform novices and attain levels of accuracy that are equivalent to experts, but more widely there is no convincing evidence yet that AI-based decision support improves clinical outcomes.[8] The development of ML could indeed have a “devastating impact” on clinical expertise if it is left to engineers rather than being viewed as an exciting challenge and opportunity for collaborating to improve the quality of diagnostic practice. The priority must be to ensure that new software tools are developed, validated, and implemented appropriately—which means that diagnostic imaging experts will continue to be needed, both to set the agenda and to introduce new tools safely.

Clinical imperatives for safe implementation

The first essential requirement is transparency.[9] Whenever an AI tool or ML algorithm is being applied, its use must be announced. It should be easy to discover details of the training dataset that was used for development and whether the tool was ever retested in an independent population. Diagnostic imaging systems should indicate to the clinician if the ML tool has been sufficiently validated to be applied in each individual case.

The precision and reproducibility of ML software is frequently exaggerated by comparing repeated measurements of the same dataset for each subject, instead of analyzing two separately acquired sets of images of the same structures. The result of using ML in any individual patient should therefore be reported with its diagnostic error or confidence intervals. Measurements should not be oversimplified (such as reporting a single number for ejection fraction without giving left ventricular volumes, when it is a ratio with limited utility). Recommendations from AI systems will only be accepted in clinical practice if they are intelligible; extreme caution is required when considering a diagnosis (“classification”) offered without explanation by a neural network.

Using AI tools in medical research

There has been a tendency during recent decades to fund experimental more than clinical research, but early hopes of major advances through better understanding of basic mechanisms of disease have not all been fulfilled. The promise of widespread health benefits from the genomic revolution, for example, has yet to be fulfilled. It has become recognized that disease may result from the net effects of complex interactions affecting common final pathways, with genetic influences, although significant, accounting for only a small proportion of the prevalence of common (rather than monogenic) diseases.[10] For further advances, patients need to be characterized by precise clinical phenotyping and their images interpreted using profound knowledge of clinical pathophysiology.

When a disease is poorly understood, then unsupervised machine learning may identify phenogroups of patients who share features that cannot be recognized using conventional imaging and statistical techniques, but it is important that the results are interpretable. Otherwise, the underlying pathophysiological mechanisms may remain obscure, and it will be impossible to develop and test treatments that will target the key steps.

An example could be heart failure with preserved ejection fraction, for which most clinical trials are widely held to have failed despite them having been performed in subjects who were recruited using suboptimal diagnostic entry criteria, such as left ventricular ejection fraction and dyspnoea. Initial studies with ML have only reinforced what we already knew.[11] Advances in treatment will remain elusive until trials are more focused, which will be possible only with the participation of clinical diagnostic specialists. AI and ML tools may provide new insights and augment their knowledge, but they will not replace their expertise.

“Cardiac imaging” as a career?

Every cardiologist—not none—should be an “imager”, in the sense that every clinician who treats patients needs to understand the mechanisms of their symptoms and pathology. Sub-specialization will have gone too far if it becomes widely accepted that the insights obtainable from detailed functional imaging performed by an experienced echocardiologist can be replaced by an automated report from a machine that has been generated by ML. That would also be an indictment of the suboptimal quality of many current diagnostic studies.

It is wrong to argue in favour of one diagnostic imaging modality rather than another—for example, for echocardiography instead of computed tomography (CT) or magnetic resonance (MR) imaging. But it is equally wrong to imply that CT or MR are better because they produce nicer tomographic images that can be obtained with less dependence on practical skills. The array of options is complementary, but CT and MR cannot replace the detailed physiological insights that echocardiography is capable of providing with precision at high frame rates. Echocardiography has reached a state of maturity, but its utility is still not fully exploited. Don’t abandon it—exploit it!

The need for (truly) intelligent cardiologists will never be replaced by (falsely) “intelligent” machines. And patients want to be cared for by fellow human beings with whom they can talk and whom they can trust. If we hand control of echocardiography to computer algorithms, we might as well abolish doctors altogether. Patients could be diagnosed by machines and algorithms, and treated according to guidelines or undergo procedures by robots—is that really what we want?

eISSN:
2734-6382
Idioma:
Inglés