Skip to main content
Cardiology Journal logoLink to Cardiology Journal
. 2020 Sep 10;27(4):404–420. doi: 10.5603/CJ.a2020.0087

Artificial intelligence and automation in valvular heart diseases

Qiang Long 1,2, Xiaofeng Ye 1,, Qiang Zhao 1,
PMCID: PMC8016001  PMID: 32567669

Abstract

Artificial intelligence (AI) is gradually changing every aspect of social life, and healthcare is no exception. The clinical procedures that were supposed to, and could previously only be handled by human experts can now be carried out by machines in a more accurate and efficient way. The coming era of big data and the advent of supercomputers provides great opportunities to the development of AI technology for the enhancement of diagnosis and clinical decision-making. This review provides an introduction to AI and highlights its applications in the clinical flow of diagnosing and treating valvular heart diseases (VHDs). More specifically, this review first introduces some key concepts and subareas in AI. Secondly, it discusses the application of AI in heart sound auscultation and medical image analysis for assistance in diagnosing VHDs. Thirdly, it introduces using AI algorithms to identify risk factors and predict mortality of cardiac surgery. This review also describes the state-of-the-art autonomous surgical robots and their roles in cardiac surgery and intervention.

Keywords: artificial intelligence, machine learning, valvular heart diseases, medical image analysis, transcathether aortic valve implantation

Introduction

Artificial intelligence (AI), which is a branch of computer science that attempts to create machines that perform tasks as though they possessed human brainpower, and this is gradually changing the landscape of the healthcare industry [1]. Clinical procedures that were supposed to, and could only be performed by human experts previously can now been carried out by machines in a more accurate and efficient way. Recently, cardiologists and radiologists have cooperated with computer scientists and have developed a fully automated echocardiography interpretation system for clinical practice [2]. Cardiovascular diseases affect 48% of the population (≥ 20 years old) and is the leading cause of death [3]. Echocardiography is an effective method to monitor the state of the heart and allows early diagnosis before the onset of symptoms. However, in primary care clinics and poorer regions, there are inadequately experienced radiologists for echocardiography interpretation, let alone patient follow-up. There is group of researchers who trained AI systems using over 14,000 echocardiograms for multiple tasks including view classification and quantification of chamber volume, ejection fraction and longitudinal strain. The AI system validated its accuracy in diagnosing several cardiac diseases after testing on over 8000 cardiograms obtained from routine clinical flow [2]. Such an AI system holds great potential for the transformation of current clinical practice models as well as the popularization of quality healthcare.

The above example is only the tip of the iceberg in the application of AI in medicine for over half a century. Since “AI” was first proposed by John MacCarthy at the Dartmouth College Conference in 1956, great efforts have been made for applying AI to almost all phases of clinical practice. In 1960s, the concept of “computer-assisted diagnosis” emerged with attempts to use mathematical formalism to interpret clinical problems, although they have little practical value, it has paved the way for subsequent expert systems [4]. In 1970s, researchers began to shift attention to studying the reasoning process of clinicians and simulate it on computer systems, which gave birth to the first generation of medical AI products, i.e. MYCIN. It attempts to diagnose infectious diseases and provide appropriate medication therapies based on over 200 rules that in the form of “if (precondition), then (conclusion or action)”. Although MYCIN showed reliability in 63% of the cases of bacteremia [5], it is more like an advanced version of a textbook which has the function of automatic retrievals with little “intelligence”. Moreover, the interactions between each rule should be defined by human experts which is extremely arduous with increasingly new knowledge added to the system. However, automatic interpretation of electrocardiograms has gained great benefits from the rule-based AI system. During the same period, a new model, i.e. the probabilistic reasoning was used to simulate the process of expert decision making [6]. Unlike the categorical mode of reasoning utilized by MYCIN, the probabilistic model assign weight to every symptom and clinical finding to indicate its possibility of occurrence for a certain disease. Based on this, new systems including the present illness program, INTERNIST, CASNET were developed [4]. With the introduction of “artificial neural network” in 1980s, machine learning began to flourish. Machine learning endows computers with the ability to learn patterns from data and perform tasks without explicit programming [7]. It has showed great power in developing robust risk prediction models, redefining patient classes and other tasks with the popularization of electronic health records and digital imaging systems [8]. In 2006, the introduction of deep learning by LeCun et al. [9] achieved unprecedented progress in image recognition and other domains, which brought a new renaissance to the development of AI and made automated medical imaging interpretation one of the hottest topics in AI. AI can generally be classified as either weak AI or strong AI. Strong AI is an intelligence construct that has the ability to understand and think as a human being. Weak AI is limited to a specific or narrow area, which is often designed to perform time-consuming tasks and analyze data in ways that humans sometimes cannot. In this review, a brief introduction to AI is provided and highlights its application in the clinical flow of diagnosing and treating valvular heart diseases, including heart sound auscultation, medical image analysis (echocardiography, cardiac computed tomography [CT] and cardiac magnetic resonance [CMR]), risk factor identification, mortality prediction and robotic surgery.

Key concepts in AI

Machine learning

In essence, machine learning is an extension to traditional statistical methods for dealing with large data sets and variables. It enables computers to learn rules and even uncover new patterns from data through a series of algorithms. Most importantly, machine learning can gradually optimize the “reasoning process” between inputs and outputs. Machine learning can be broadly classified into three classes: supervised learning, unsupervised learning and reinforcement learning.

In supervised learning, machines learn the mapping relations between input variables and labeled outcomes and are able to predict outcomes according to inputs. For example, after training with echocardiograms that are labeled with given disease categories, machines are able to assign these disease labels to new echocardiograms. Common computational approaches in supervised learning include artificial neural network (ANN), support vector machine, k-nearest neighbor, naive Bayesian model and decision tree model, among others. Unsupervised learning aims to uncover hidden structures in unlabeled data and classify it into separate categories, which means machines are only trained with input variables and automatically find potential classification rules. For example, with unsupervised learning, machines uncovered three cardiac phenotypes in patients diagnosed with type 2 diabetes mellitus according to their echocardiograms [10]. Common computational approaches include hierarchical clustering, k-means clustering and principle component analysis. Reinforcement learning signifies that machines learn strategies which can obtain the maximum reward through interaction with the environment or outcomes. Unlike the static mode of supervised or unsupervised learning, it is a dynamic learning process. Q-learning, an example of reinforcement learning, has been utilized in clinical trials of lung cancer to find the optimal individualized therapies [11].

Artificial neural network and deep learning

Artificial neural network is composed of multiple interconnected artificial neurons which mimic the biological brain. As mentioned above, ANN is actually a computational model. ANN learns from large amounts of data and continuously optimizes the weight of each connection between neurons. Figure 1B shows a typical ANN, which is also called multilayer perceptron, is composed of three layers: input layer, hidden layer and output layer. Each layer contains multiple neurons that are responsible for different functions. The input layer receives numerical variables from data, each value in first layer neurons combines with the weight of the connection, which is then propagated to a hidden layer where the values are integrated, while an output layer generates the final value which represents an outcome. Each neuron in the hidden layer can also set up a threshold to determine whether the information should propagate to the next layer. Now suppose that a database that includes patient data (such as age, body mass index, laboratory results and images results) prior to cardiac surgery and operation results (alive or dead), a mortality prediction model can be built with ANN. Patient variables are input into the neurons in the first layer, the outputs are set as 1 for alive and 0 for dead. The weight of each connection and the excitation threshold of each neuron in the hidden layer are gradually optimized by learning from the data. In this way, after trained with sufficient cases, the ANN model can be accurate enough to predict surgical outcome in new patients.

Figure 1.

Figure 1

Schematic diagram of the evolvement of artificial neural network. Perceptron, also called single layer neural network, was proposed by Rosenblatt in 1958 (A), classic artificial neural network, also known as double layer neural network or multilayer perceptron (B), deep neural network (C).

It can be imagined that, with more hidden layers, the computational results will be more accurate, but this meanwhile requires greater computational power using more training data. Thanks to the rapid development of computer processors and arrival of the era of big data, and also the introduction of “pre-training process”, “fine-tuning technique” by Hinton et al. [12] which decreases the training time substantially, multilayer ANN which is also known as “deep learning” becomes a reality. It now represents the most advanced AI technology. The biggest advantage of deep neural network is that it allows the computing model to automatically extract the features of the data by abstracting the data layer by layer and learning the representative and discriminant features.

Computer vision

Computer vision is a main research field in AI, which simulates the human vision with imaging systems and computers. The main tasks involved in computer vision can be divided into three levels [13]. The basic level includes image acquiring and processing techniques such as noise filtering, image enhancement, image segmentation, pattern recognition. For middle level computer vision, computers should be able to draw conclusions and make decisions based on the information obtained from the basic level. For the top level, computers have the capability to “think” and can understand images like humans do. For now, the first two levels of computer vision are mostly applied to the analysis of medical images and assist in diagnosis. As described below, many algorithms and commercial software have been developed for automating the analysis of medical images, which are more accurate and efficient compared to human endeavour.

Expert system

Expert systems are in their earliest application of AI in medicine, and will keep playing an important role in healthcare. In brief, expert systems are a computer program that simulates human experts in solving problems [14]. It is also called a knowledge-based system because it contains a large amount of knowledge and experience from experts. A typical medical expert system is composed of two modules: knowledge base and control system, the latter can be further divided into a human-computer interactive component, an explanation component, a knowledge acquisition component and an inference engine. The knowledge base is composed of medical knowledge from experts, case-specific knowledge from patients and intermediate results from the reasoning process [14].

Application of AI and automation in diagnosis of valvular heart diseases

AI-assisted “cardiac auscultation”

In spite of the diverse alternatives of imaging tools being the major diagnostic approach to valvular heart diseases (VHDs), physical examination, is the cornerstone of clinical diagnosis and should be the primary screening method for VHDs [15]. Common physical examinations for VHDs include general inspection, pulse palpation, percussion of heart boarder and heart sound auscultation, etc. [16]. Among these, auscultation plays a key role of diagnosing VHDs.

Valvular heart diseases manifest as heart murmurs or/and extra heart sound before the stress of hemodynamic changes causing other signs and symptoms including dyspnea, fatigue, angina, cough and hemoptysis. While cardiac auscultation provides significant diagnostic and prognostic information for cardiac disease, it is not an easy skill to master, especially when innocent murmurs are confronted. Although experienced clinicians have reached an accuracy over 90% in identifying innocent murmurs from pathologic murmurs, less-experienced residents and primary care physicians perform less than satisfactorily [17]. In addition, auscultation is a highly subjective process which may cause bias when evaluating the intensity, location and shape of murmurs. For over a century, tools for cardiac auscultation is a mechanical stethoscope, however, it can neither store nor play back sounds [15]. Thanks to the invention of electronic (digital) stethoscope, the forgoing problems have been well solved, more than that, it provides clinicians with a handy way of “seeing” the heart sound through phonocardiogram. AI-assisted cardiac auscultation in practice refers to the auto-interpretation of phonocardiogram, which belongs to the domain of signal processing. Key steps involved in heart sound analysis could be summarized as segmentation, feature extraction and classification as shown in Figure 2 [18]. Each step is fulfilled through a multitude of algorithms with the ultimate goal of precisely identifying the pathological events underlying heart sounds.

Figure 2.

Figure 2

Process of artificial intelligence-assisted auscultation. A normal phonocardiogram (A), heart sound segmentation (B), feature extraction (C), heart sound classification by deep neural network (D).

Heart sounds segmentation

Segmentation aims to locate the fundamental elements including the first heart sound (S1), systolic period, the second heart sound (S2) and diastolic period in each cardiac cycle. Training machines to think like human so as to solve problems is to mimic the thinking process of human brain to a certain extent. When interpreting heart sounds, A human expert would firstly locate the two fundamental heart sounds then discriminate S1 from S2 by its pitch, intensity and duration, finally the systolic and diastolic regions are determined, as is the process with computational analysis, but in a more logical and mathematical way.

In the early exploration of heart sound segmentation, electrocardiograph or/and carotid pulse was/were obtained concurrently with phonocardiogram [19, 20]. Transformation equation g(n) was used to compute the smooth energy curve of electrocardiograph signal, then the peaks in g(n) was determined as the beginning of S1 and the systolic period, the dicrotic notch of carotid pulse curve was recognized as the beginning of S2 [19]. This approach apparently is impractical in routine clinical work. Envelop-based (or amplitude threshold-based) segmentation algorithm was first introduced by Liang et al. [21] to locate S1 and S2. Envelop of a signal signifies the smooth curves outlining its upper or lower extremes [22]. The envelop of the original heart sound waveform was first extracted using specific transformation equations, then a threshold value was selected to filter noise and low intensity signals, the peaks exceed the threshold value were recognized as the S1 and S2, S1 was differentiated form S2 based on an assumption that the systolic period was shorter than the diastolic period [21]. A well-fitted extraction equation would allow only two peaks to exceed the threshold when interpreting the normal heart sound, therefore identification accuracy depends largely on an envelope extraction equation. Diverse methods are used to calculate the envelope were created including normalized average Shan-non energy [21], single-DOF analytical model [23], wavelet decomposition method [24], Viola integral approach [25], and others. However, the envelope-based segmentation method often leads to problems like weak peaks or extra peaks which reduce accuracy of the results.

The application of machine learning algorithms to heart sound segmentation based on features substantially improves the segmentation results (see feature extraction below). Hidden Markov Model (HMM) is the most widely used in the task [2628]. To train the HMM, adequate samples pre-labeled with the accurate location of S1, S2, systolic and diastolic period are required at the output end; sequences of feature vectors extracted from original phonocardiogram or transformed envelope are used as the observation end of HMM. Recently, utilization of deep learning achieves even higher precision compared to other classification methods in heart sound segmentation [29, 30].

Feature extraction and feature selection

This step aims to select and extract discriminative features either for more precise segmentation of heart sound or for following the disease classification step. Feature means the characteristic of an object by which the human brain recognizes and distinguishes it automatically. The concept of feature is similar to “variable” in regression analysis. Features that can be recognized by machines tend to be presented in the form of numbers or symbols. A human expert draws physiological or pathological information from heart sound through features including heart rate, heart rhythm, timing and shape of the murmur, pitch of the heart sound, and extra heart sound among others. As for phonocardiogram, the features are based on time-domain, frequency-domain, and time-frequency complex domain [31]. Time-domain features include intervals (interval of S1 and S2, systolic intervals, diastolic intervals, the ratio of each component to the cardiac cycle etc.) and amplitude (mean absolute amplitude of S1 and S2 interval). Frequency-domain features refer to the power spectrum of each heart sound component across frequency bands [29]. Theoretically the more input of features into machine training, the better the classification performance it will achieve. In practice, with the number of training samples being set, the classification performance will drop off when the number of feature inputs exceeds a certain value. In order to exclude the redundant features and improve the classification efficiency, feature selection is often required [32].

Classification and detection of VHDs

This step aims to classify the phonocardiograms into cardiac disease categories using suitable classifiers. Classifier in machine learning refers to the algorithm that learns to assign labels to testing data from training data. This learning process can be supervised or unsupervised. Briefly speaking, supervised learning assigns given labels to data while unsupervised learning seeks labels that could be assigned to data. In the present case, labels are given disease categories like mitral stenosis, mitral regurgitation, aortic stenosis, etc. therefore, supervised learning algorithms/classifiers are utilized. The common classifiers used in heart sound classification include a support vector machine [24, 3336], neural networks [37, 38], HMM [39], etc. After a careful combination of algorithms used in segmentation, feature extraction and classification, previous studies showed promising and inspiring results in detecting VHDs from phonocardiograms (Table 1). For instance, an intelligent diagnostic system developed by Sun [24] could discriminate aortic regurgitation, mitral regurgitation, pulmonary stenosis with accuracy of 98.9%, 98.4% and 98.7%, respectively. Thompson et al. [17] utilized a murmur detection algorithm developed by CSD labs to distinguish no murmurs and innocent murmurs from pathologic murmurs. 3180 phonocardiograms recorded at five different chest location from 603 cases were tested, the algorithm had good sensitivity, specificity and accuracy in detecting a pathologic murmur, which are 93%, 81% and 88%, respectively. The algorithm showed the highest accuracy (90%) using recordings from the left upper and lower sternal boarder. However, it was unable to analyze the recordings with precision which had a high heart rate and low signal quality [17].

Table 1.

Sensitivity/specificity or accuracy of artificial intelligence-assisted auscultation system in detecting valvular heart diseases.

Reference Year Algorithm Testing samples Normal (%) AS (%) AR (%) AS+AR (%) MS (%) MR (%) MS+MR (%) PS (%) PR (%)
Fahad et al. [39] 2018 HMM 80 abnormal HSs and 150 normal HSs 100/99.33 92/100 96/98.54 93.33/99.50
Sun [24] 2015 SVM 588 abnormal HSs from 49 patients, 325 normal HSs from healthy individuals 99.8 98.8 98.4 98.7
Zhang et al. [33] 2014 SVM 40 abnormal HSs and 75 normal HSs 98.67 85 95
Safara et al. [34] 2013 SVM 43 abnormal HSs and 16 normal HSs 98/100 97.44/98.40 100/100 95.65/98.31
Kwak et al. [35] 2012 HMM 53 abnormal HSs, 80 normal HSs 99 67 67 67 83 67 80
Uğuz [37] 2012 ANN 40 abnormal HSs, 20 normal HSs 95 95 95
Kumar et al. [36] 2010 SVM 81 abnormal HSs from 51 patients 93.02/91.29 94.78/93.98 95.20/96.33 100/100 93.21/95.80
Olmez et al. [38] 2003 ANN 240 abnormal HSs from 20 patients, 48 normal HSs from 4 healthy individuals 100 97.9 100 100 100 97.9

If more than one testing methods were utilized in the study, only the method with highest sensitivity/specificity or accuracy is presented; ANN — artificial neural network; AR — aortic regurgitation; AS — aortic stenosis; HMM — hidden Markov model; HS — heart sound; MR — mitral regurgitation; MS — mitral stenosis; PR — pulmonary regurgitation; PS — pulmonary stenosis; SVM — support vector machine

AI-assisted interpretation of echocardiography

Although heart sounds auscultation is convenient, cost-effective and a quick way to diagnose VHDs, it only provides qualitative diagnostic information. To further confirm the diagnosis of VHDs as well as to assess the etiology, severity, ventricular responses and prognosis, imaging tools including echocardiography, CMR, multi-slice CT (MSCT) or even cardiac catheterization are indispensable [40].

Echocardiography is the “heart” of cardiology. It is the preferred method of diagnosing and guiding the treatment of VHDs as well as other cardiac diseases. However, echocardiographic examination is a time-consuming process which usually takes hours from inspection, analysis to formal report which make it impractical in emergency settings. Moreover, typical echocardiographic examination produces large amounts of data including images from multiple views and up to 70 videos which would cause cognitive overload and be prone to error. In addition, due to the characteristic of multiple views in examination, the problems of intra-observer and inter-observer variability inevitably arise. The application of AI and automation in echocardiography would a allow consistent, quick and accurate measurement which promises a more accurate diagnosis and improved patient care.

Image segmentation-valve leaflet detection and tracking

Segmentation and recognition of anatomical structures from original medical images are preprocessing steps for subsequent quantitative analysis and diagnosis. Morphological characteristics and motion pattern of the heart valves are key information to diagnose VHDs and to assist surgical valve repair or percutaneous intervention. While many studies have sought the approaches to segment heart chambers and detect endocardial boarder which were reviewed by Noble and Boukerroui [41], relatively few studies of the segmentation of heart valves and annulus have been published. The reasons for difficulty in segmenting and tracking the heart valves in echocardiographic sequences can be summarized as follows: (1) poor image quality due to low frame rate, speckle noise and artifact which may result in missing boundaries; (2) lack of features to discriminate heart valve from adjacent myocardium for those structures have similar intensity and texture; (3) fast and irregular valve motion making it hard to establish correspondence between frames. Reported approaches for valve segmentation including active contour models [42, 43], thin tissue detector combined with graph cut [44], outlier detection method [45, 46], J-spline method [47], multi-atlas joint label fusion [48], trajectory spectrum learning algorithm [49] and neural network [50].

Active contour models (ACM) is the first and most widely used segmentation algorithm for the heart valve [51, 52]. The segmentation process is based on ACM and is initiated by manual placement of contour close to the target, which means it is not fully automated. Zhou et al. [45] formulated the problems of valve detection and tracking as outliner detection in the low-rank representation based on different motion patterns between leaflets and the heart wall, which is fully automated requiring no user interaction [46]. Recently, there is a growing trend of applying machine learning to valve segmentation. This approach shifts the manual input to a training phase which makes segmentation fully automated [50]. Ionasec et al. [49] utilized spectrum learning algorithm and boundary detectors to locate and refine mitral leaflets based on anatomical landmarks in four-dimensional (4D) transesophageal echocardiograms and CT which allows precise morphological and functional quantification, and a large amount of data as well as high computational power were required to train the model. Recently, a novel UNet architecture based on a convolutional neural network was utilized by Costa et al. [50] to segment mitral valves in PLAX and A4C views. The biggest strength of UNet architecture is that it does not require a large training dataset to produce accurate results.

Automated quantitative analysis

Precise and reliable quantitative analysis of stenosis and regurgitation is crucial to severity assessment, prognosis prediction and to evaluate whether surgery or intervention is needed. Proximal isovelocity surface area (PISA) method is widely used in measuring mitral valve orifice area and calculating the regurgitant volume [53]. Conventional PISA method of two-dimensional echocardiography is based on the assumption that the shape of the proximal flow convergence region is hemispheric, which is not quite true [54]. The development of three-dimensional (3D) color Doppler echocardiography enables direct measurement of PISA without assumptions, thereby reducing errors [54]. However, it is not indicated for routine clinical use because it is a time-consuming procedure, thus in desperate need of an automated algorithm. Grady et al. [55] proposed the first automated PISA measurement system based on a random walker algorithm. The system was initiated by the manual input of two points (one at the valve annulus and one at the coaptation site), followed by automated segmentation of valve annulus and the isovelocity region. The segmentation results were then used to generate 3D meshes for computing PISA. In vitro experiments validated its accuracy in measuring PISA, effective regurgitant surface area (EROA) and regurgitant volume. Further in vivo experiments showed that measurement of EROA on patients with magnetic resonance (MR) by automated algorithm was significantly correlated with manual measurement of vena contracta [55]. Cobey et al. [56] proposed another novel automated method based on Halcon HDevelop Machine Vision. 3D images of PISA were first sliced into 2 mm-thick sequential cuts, the algorithm traced the boundary of each slice and the arc lengths between each slice from which the 3D surface area was generated and computed. Another method of quantitatively assessing MR is to use regurgitant fraction which is calculated from mitral inflow and stroke volume [53]. Wang et al. [57] proposed a novel automated system for estimating mitral inflow and aortic outflow. The system first detected the left ventricular wall, mitral annulus and left ventricular outflow tract using marginal space learning and placed the measurement plane. The 3D motions of these structures were tracked through a whole cardiac circle to construct and adjust the measurement plane, then the volume of mitral inflow and left ventricular outflow was computed by aggregating color flow values in the 3D space [57].

In addition to these pioneering studies, several commercial software have been developed including Mitral Valve Quantification (Philips Medical Imaging, Andover, MA) [58], eSie Valve (Siemens Healthcare, Mountain View, CA, USA) [59], Mitral Valve Navigator (Philips Medical Systems) [60], Auto Valve Analysis (Siemens; California, USA) [61], eSie PISA Volume Analysis (Siemens Medical Solutions USA, Inc., Mountain View, CA) [62] which aim to automate the quantitative analysis of 3D echocardiography. These approaches can reduce the measurement time substantially and provide more accurate and reproducible results. The aforementioned automated algorithms and commercial software have validated their reliability in quantitatively measuring aortic and mitral valve apparatus parameters and regurgitation volume [54, 63, 64] as shown in Table 2.

Table 2.

Automated software and algorithms in quantification analysis of three-dimensional echocardiography.

Reference Year Software/algorithm Function Reference method Samples ICC/intergroup difference between software/ algorithm and reference method) Intra-observer variability (ICC) Inter-observer variability (ICC)
Mediratta et al. [58] 2017 Mitral Valve Quantification (QLAB version 9.0.1; Philips Medical Imaging, Andover, MA; USA) For measurement of aortic annulus area, perimeter and diameter Computed tomography 52 patients ICC of annulus area, perimeter, area-derived diameter, perimeter-derived diameter: 0.91, 0.90, 0.91, 0.89 0.83, 0.85, 0.81, 0.85, respectively 0.70, 0.72, 0.72, 0.71, respectively
Aquila et al. [59] 2016 eSie Valve Software (Siemens Healthcare, Mountain View, CA, USA) For measurement of mitral valve parameters Manual measurement 61 patients ICC of inter-commissure diameter, mitral annular area, anterior leaflet length, posterior leaflet length: 0.84, 0.94, 0.83, 0.67 0.999, 0.992, 0.984, 0.937, respectively 0.997, 0.957, 0. 963, 0.936, respectively
Kagiyama et al. [60] 2015 Mitral Valve Navigator (Philips Medical Systems) For measurement of mitral valve parameters Mitral Valve Quantification (Philips Medical Systems) 74 patients Cronbach’s alpha of commissural width, anteroposterior diameter, annulus area: 0.902, 0.934, 0.938 Not available 0.945, 0.959, 0.961, respectively
Calleja et al. [61] 2013 Auto Valve Analysis (Siemens; California, USA) For measurement of aortic annulus diameter, leaflet anatomy and coronary ostium height Computed tomography 40 patients Intergroup difference (p value) of left, non-coronary, right leaflet: 0.76, 0.06, 0.29; left and right coronary height: 0.003, 0.001 Annulus diameters at sagittal and coronal view: 0.97, 0.92 Annulus diameters at sagittal and coronal view: 0.91, 0.93
Agustin et al. [62] 2013 eSie PISA Volume Analysis (Siemens Medical Solutions USA, Inc., Mountain View, CA) For measurement of EROA and RVol Manual measurement by 3D planimetry 90 patients ICC of EROA, RVol: 0.97, 0.97 Not available Not available
Choi et al. [63] 2015 Based on Wang [71] For measurement of aortic RVol CMR 32 patients ICC of aortic RVol: 0.93, AR severity: k = 0.94 0.98 0.96
Thavendiranathan et al. [64] 2013 Based on Grady [69] For measurement of mitral RVol CMR 35 patients ICC of Integrated PISA-derived RVol, flow volume-derived RVol: 0.92, 0.91 0.99, 0.93, respectively 0.95, 0.96, respectively

Not all results are listed in the table; 3D — three-dimensional; AR — aortic regurgitation; CMR — cardiac magnetic resonance; EROA — effective regurgitant orifice area; ICC — interclass correlation coefficient; PISA — proximal isovelocity surface area; RVol — regurgitant volume

AI-assisted interpretation of cardiac CT images

Computed tomography is not employed as the preferred diagnostic method of VHD. It generally serves as a complementary role when echocardiography is insufficient or inclusive. Nevertheless, cardiac CT has prominent advantages in the evaluation of valve calcification and annulus geometry. It is the golden standard for annulus sizing and an indispensable tool in preoperative planning of transcatheter aortic valve replacement (TAVR). Reliable measurement of aortic valve annulus size, aortic root dimension and the height of coronary ostia are crucial for appropriate transcatheter valve prosthesis selection and intraprocedural valve positioning which may improve success rate and reduce postoperative complications. In recent years, advances in 3D image techniques have enabled reconstruction of patient-specific models of aortic valve apparatus. Several automated algorithms and commercial software for aortic valve segmentation and quantification have been developed for preoperative planning of transcathether aortic valve implantation (TAVI).

Patient-specific 3D modeling of aortic valve

Ionasec et al. [49, 65] proposed the first dynamic patient-specific aortic valve model from 4D CT images. A generic physiological model which can represent aortic valve and its pathological variations was constructed first, patient-specific parameters of the model were estimated from volumetric sequences by trajectory spectrum learning, marginal space learning and discriminative learning. Their study was further improved with shape forest to constrain the classical statistical shape model [66]. Waechter et al. [67] extracted the aortic valve geometry from CT images using model-based segmentation and applies pattern search method to detect the coronary ostia. The whole heart geometry (including heart chambers and great vessels) was first roughly modeled, a generic aortic mesh model was established and boundary detector were trained by annotated images aiming to build more detailed models of the aortic valve and root, the coronary ostia was then detected on the surface of the aortic root. Segmentation results allowed a series of clinical measurements including the diameters of annulus and the distance between ostia and aortic valve, etc. Considering that VHDs often involve multivalvular lesion which require joint assessment, Grbic et al. [68] proposed an integrated model for quantification of all heart valves from 4D CT images based on marginal space learning and multi-linear shape models. Another study of Grbic et al. [69] extracted both the volumetric model of the aortic valve and calcification, which is closely correlated with postoperative regurgitation [70].

Landmarks detection and quantitative measurement

A number of parameters should be measured with precision for surgical planning before TAVI. These include annulus diameter, annulus area, angulation of annulus plane and the distance from annulus to coronary ostia to name a few. The annulus diameter is crucial for selection of the appropriate valve prosthesis. Size mismatching may either cause post-operative perivalvular regurgitation or annulus rupture. A short distance between aortic annulus and coronary ostia indicates elevated risk of coronary obstruction after valve deployment. Annulus plane angulation determines the position of an X-ray tube C-arm during operations to achieve an optimal view of valve delivery [71].

Accurate segmentation of the aorta, aortic valve apparatus and detection of anatomic landmarks (coronary ostium, aortic commissures and aortic hinges) is the prerequisite for reliable measurement of the above parameters. Zheng et al. [72] proposed a robust hierarchical approach by first segmenting the global aorta using marginal space learning from which the position of the anatomical landmarks can be roughly inferred, followed by using specific landmark detectors to refine each landmark. Elattar et al. [73, 74] used thresholding and connected component analysis to detect the region of interest, from which aortic root was extracted using 3D normalized cut. Two coronary ostia and three valve hinge points were then detected on the surface of aortic root by intensity projection map and Gaussian curvature map. Lalys et al. [75] used a hybrid approach which integrated thresholding, model-based method, statistical-based method and a 3D contour model in the procedure, however, a user-specific point is needed to define the volume of interest. Recently, colonial walk algorithm, which is a machine learning method, was utilized by Al et al. [76] to automatically localize the landmarks. Several commercial software including three mensio valves (3mensio Medical Imaging BV, the Netherlands) [77, 78, 81], Syngo (Siemens Healthcare, Erlangen, Germany) [79], IntelliSpace Portal (Philips Medical Systems, Cleveland, OH) [80], have been available for routine clinical use. Together with the aforementioned algorithms, these pioneering techniques showed reliable aortic annulus measurements as shown in Table 3.

Table 3.

Automated software and algorithms in quantification analysis of aortic valve apparatus.

Reference Year Software/algorithm Reference method Samples Correlation/bias Inter-observer variability Intra-observer variability
Dankerl et al. [79] 2017 Syngo, CT TAVI Planning (version VB10A, Siemens Healthcare, Erlangen, Germany) Manual measurement (software: NA) 30 patients Distance from annular plane to coronary ostia: r = 0.97 (right); r = 0.93 (left)
Annulus area: r = 0.89
Annulus diameter: r = 0.89 (maximum diameter); r = 0.82 (minimum diameter)
NA NA
Guez et al. [80] 2017 Intellispace Portal (Philips Medical Systems, Cleveland, OH) Manual measurement using 3D Brilliance workstation (Philips Healthcare) 74 patients Annular diameter in systole: r = 0.85 (minimum), r = 0.87 (maximum diameter)
Annular diameter in diastole: r = 0.89 (for both minimum and maximum diameter)
Annular area: r = 0.93 (systole), r = 0.94 (systole)
NA NA
Watanabe et al. [77] 2013 3mensio Valves (version 5.1, 3mensio Medical Imaging BV, the Netherlands) Manual measurement using dedicated workstation (EBW, Philips Healthcare) 105 patients Average bias: mean annulus diameter: 1.15 mm, left coronary ostium height: 1.61 mm, right coronary ostium height: 3.45 mm ICC: 0.93 to 0.98 ICC: 0.94 to 0.98
Delgado et al. [78] 2011 3mensio Valves (version 4.1.sp1, Medical Imaging BV, Bilthoven, The Netherlands) Manual measurement using Vitrea2 software (Vital Images, Minneapolis, MN) 90 patients ICC of annulus diameter, LVOT diameter, sinus of Valsalva height, coronary ostium height: 0.97 to 0.99 ICC: 0.90 to 0.99 ICC: 0.95 to 0.99
Lalys et al. [75] 2019 Hybrid approach: model-based method, active contour model, atlas-based registration, learning-based detection Manual measurement (software: NA) 50 patients Detection bias of aortic leaflets 1.96 ± 0.84 mm, 2.34 ± 0.78 mm and 1.23 ± 1.12 mm (right, left and non-coronary ostia)
Measurement bias: annulus diameter: 1.31 ± 0.75 mm, coronary ostia heights 1.96 ± 0.87 mm (right), 1.80 ± 0.74 mm (left)
NA NA
Al et al. [76] 2018 Colonial walk Manual measurement (software: NA) 71 patients Mean detection bias of hinge points, commissure points and coronary ostia: 2.09 ± 1.20 mm
Measurement bias: aortic annulus diameter: 0.94 ± 0.83 mm, annulus to ostia distance: 0.98 ± 0.90 mm
NA NA
Elattar et al. [73] 2016 3D normalized cut and intensity map Manual measurement (software: NA) 40 CTA images landmark segmentation bias: 2.38 ± 1.56 mm
Measurement: ICC of annulus radius: 0.84, annulus to ostia distance: 0.73
NA NA
Elattar et al. [74] 2014 3D normalized cut method Manual delineation 20 patients Dice coefficient of aortic root segmentation: 0.95 ± 0.03 Dice coefficient: 0.951 ± 0.03 NA
Zheng et al. [72] 2012 Marginal space learning and optimization-cased method Manual measurement (software: NA) 319 C-arm CT volume images Segmentation bias: aortic hinges 2.09 ± 1.18 mm, coronary ostia 2.07 ± 1.53 mm, aortic commissure: 2.17 ± 1.88 mm
Measurement bias: distance from annulus plane to coronary ostia 2.10 ± 1.86 mm (left), 2.51 ± 2.03 mm (right)
NA NA

Not all results are listed in the table; 3D — three-dimensional; CTA — cardiac magnetic resonance; CT — computed tomography; ICC — interclass correlation coefficient; LVOT — left ventricular outflow tract; NA — not available

A comparative study between 3mensio, Intelli-Space Portal (version 7.0), IntelliSpace Portal (version 9.0) and manual measurement concerning annulus parameters, time-cost and reproducibility was performed [81]. Results showed that IntelliSpace Portal (version 9.0) allowed the fastest and most reproducible measurements, parameters derived from either method can be used interchangeably in prosthesis sizing. Moreover, Samim et al. [71] performed a prospective cohort study to compare angiography and MSCT (using 3mensio) in predicting the annulus plane and choosing the C-arm position [71]. The study included 35 patients in an angiography cohort and 36 patients in an MDCT cohort. The utilization of MDCT was associated with a significant reduction of prosthesis implanting time, radiation exposure and contrast delivery. A reduction of postoperative complications and 30-day mortality was observed in patients in the MDCT cohort.

AI-assisted interpretation of CMR

Cardiac magnetic resonance is the second-line technique in the assessment of VHDs. A standard CMR scan takes about an hour followed by tedious image processing which is both time and labor consuming. However, it is the golden standard in the assessment of cardiac morphology, volume and function due to its high spatial resolution [82]. MR cine imaging and phase-contrast velocity imaging are gaining increasing importance in studying valvular function. Particularly, CMR have advantages over other imaging tools in the quantification of regurgitation when the regurgitant jet is highly eccentric.

Automated cardiac chamber segmentation and cardiac volume measurement have been widely studied in MR images as reviewed by Bernard et al. [83] and Petitjean et al. [84]. Several large CMR image datasets (Sunnybrook, STACOM, MICCAI RV and kaggle), which are free to access, have been released in the last decade in conjunction with international challenges to automatically segment left ventricle, right ventricle, end-diastolic volume and end-systolic volume [83]. Commercial software such as SuiteHEART® (Neosoft, Pewaukee, Wisconsin, USA) which allows automated quantification of biventricular volumes and function is now available [85]. The regurgitant volume and fraction in isolated aortic or mitral regurgitation can be calculated from left ventricle and right ventricle stroke volume [82], however, to date, the aforementioned automated methods have not applied in quantification of valve regurgitation. AI-assisted CMR which allows automated diagnosis of VHDs is underexplored. However, several pioneering studies have achieved many valuable results. Fries et al. [86] developed a novel supervised deep learning model for aortic valve malformation classification using unlabeled MR images. 570 patients were classified as bicuspid aortic valve from a cohort of 9230 patients from the UK biobank. These individuals showed a significant lower major adverse cardiac event-free survival rate compared to individuals with a normal aortic valve.

Risk factor identification and in-hospital mortality prediction of cardiac surgery

Preoperative assessment of surgical risk is an important procedure in cardiac surgery which guides the selection of surgery, intervention or nonsurgical treatment. Great progress has been made in risk factor identification and mortality prediction. Risk score models including EuroSCORE, STS score and ACEF scores have been widely used [87]. The linear regression model is the most widely used in analyzing a correlation between risk variables and cardiac adverse events. However, there is a growing trend of utilizing machine learning methods in developing prediction models which outperform traditional scoring systems.

Nilsson et al. [88] were among the first to use an ANN model to identify risk factors and predict mortality in cardiac surgery. 72 risk factors were evaluated from 18,362 patients, 34 of the factors were identified as relevant to mortality. Receiver operating characteristic (ROC) area of the ANN model for mortality prediction was significantly larger than logistic EuroSCORE model (0.81 vs. 0.79). For isolated valve surgery (with or without coronary artery bypass grafting), the ROC area of the ANN model was 0.76 vs. 0.72 of the logistic models. Celi et al. [89] achieved an even larger ROC area using ANN, Bayesian network and logistic regression (0.941, 0.931, 0.854) compared to EuroSCORE (0.648). Allyn et al. [90] utilized a novel ensemble machine learning method which integrates the results from four isolated machine learning algorithms to predict cardiac surgery mortality. 6250 patients were enrolled in the study, Chi-square filtering was utilized to extract relevant variables. Results showed that the ensemble machine learning model had a significantly stronger discriminatory power for operative mortality than EuroSCORE II (ROC area: 0.795 vs. 0.737). Recently, Hernandez-Suarez firstly applied four machine learning algorithms in mortality prediction after TAVI [91]. 10,883 patients were enrolled in the study. Logistic regression, ANN, Naïve bayes and random forest which are the four top supervised learning algorithms all showed good discriminative performance. The best prediction model obtained by logistic algorithm (ROC area: 0.92) have close discriminative power with state-of-the-art National Inpatient Sample TAVR score model.

Moreover, ANN and machine learning algorithm have been applied in other fields such as prediction of length of stay in intensive care unit after cardiac surgery [92], post-operative complications [93], both short-term and long-term mortality after heart transplantation [94]. However, the downside of current studies is that few studies have been dedicated to isolated cardiac surgery procedures which is partially due to a lack of sufficient samples to feed machine training.

Intelligent cardiac operating room: Towards autonomous robotic surgery

The introduction of minimally invasive strategies and a surgical robot in cardiac surgery promises quicker recovery and less postoperative complications and mortality. While having many superiorities over human hands such as tremor resistance and scalable motion, current robotic systems are merely teleoperated devices under human control which entirely possess no autonomy. Since automation has gained great success in other robotic fields which have increased safety, accuracy and efficiency, it is reasonable to assume that the same benefit will be gained by developing autonomous surgical robots. To achieve the level of fully “autonomous”, the surgical robot should possess the ability to “see”, “think” and “act”. “see” refers to perception of the surgical field and itself through sensors. “think” is the process of receiving information and calculating the future status that it needs to achieve in the following “act” [95]. Recently, European research council launch the Autonomous Robotic Surgery (ARS) project aiming to developing a unified framework for the autonomous execution of robotic tasks [96]. Main research objectives include establishment of global action model by analyzing current robotic surgical data, patient specific intervention models, design of controllers, perception of overall surgical situations, and assessment of the surgical robot capability.

Though at an experimental stage, the feasibility of autonomous robots performing simple surgical tasks like suture and knot typing have been demonstrated. Penesar et al. [97] developed a novel Smart Tissue Autonomous Robot (STAR) which was able to perform linear continuous suture, it successfully completed in-vivo end-to-end anastomosis of porcine small intestine with few suturing mistakes, no complications were observed in 7 day follow-up. Although the STAR system realized autonomous suture, the step of knot typing remained manual. In the context of endoscopic surgery or robot-assisted surgery, the task of knot typing is cumbersome which requires manipulation of many subtle movements in a confined space [98]. Through training a recurrent neural network on 3D loop trajectories generated by human surgeon, the Endoscopic Partial-Autonomous Robot (EndoPAR) was able to accurately perform the winding portion of a knot-tying task [99]. To date however, robots are unable to fully perform autonomous knot tying.

Transcatheter therapy has been accepted as an alternative to traditional cardiac valvular surgery especially in patients with high surgical risk in recent years. Catheters are inserted either from peripheral vessel or the cardiac apex to deploy valve prothesis or occlusion device. In either case, the catheters need to be precisely navigated to the intervention site which is a very challenging task in a beating heart. Inspired by wall following, which is used by thigmotactic animals to locate and navigate themselves in low-visibility environments, Fagogenis et al. [100] developed an autonomous intracardiac catheter navigation system. With hybrid imaging and touch sensor installed on the catheter tip, which can provide clear images of what it has touched and identify it as blood, valve or myocardium. The catheter created continuous low-force contact with the surrounding tissue and followed the cardiac wall to achieve autonomous navigation. With the navigation system, authors designed a robotic catheter, inserted from the cardiac apex, which can autonomously navigate to the aortic valve and deploy an occlusion device into the leak site on a porcine model. The in-vivo animal experiments demonstrate that an autonomous robot catheter was non-inferior to human experts [100].

Conclusions

Artificial intelligence is changing the landscape of healthcare. The inclusion of AI and automated algorithms in medical image analysis are very promising for they require less measurement time and meanwhile provide more accurate and reproducible results, which make them ideal helpers in a busy clinical flow. Machine learning methods can make full use of patient data and build more powerful prediction models of cardiac surgery compared to traditional statistical approaches. Moreover, the autonomous surgical robot, although in its infancy, holds great promise for improved safety and efficiency in cardiac surgery and intervention. In general, AI has enhanced the ability of diagnosis and clinical decision-making in VHDs and taken another step towards precision in medicine.

Footnotes

Conflict of interest: None declared

References

  • 1.Topol EJ. High-performance medicine: the convergence of human and artificial intelligence. Nat Med. 2019;25(1):44–56. doi: 10.1038/s41591-018-0300-7. [DOI] [PubMed] [Google Scholar]
  • 2.Zhang J, Gajjala S, Agrawal P, et al. Fully automated echocardiogram interpretation in clinical practice. Circulation. 2018;138(16):1623–1635. doi: 10.1161/CIRCULATIONA-HA.118.034338. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Benjamin EJ, Muntner P, Alonso A, et al. Heart Disease and Stroke Statistics — 2019 Update: A Report From the American Heart Association. Circulation. 2019;139 doi: 10.1161/CIR.0000000000000659. [DOI] [PubMed] [Google Scholar]
  • 4.Schwartz WB, Patil RS, Szolovits P. Artificial intelligence in medicine. Where do we stand? N Engl J Med. 1987;316(11):685–688. doi: 10.1056/NEJM198703123161109. [DOI] [PubMed] [Google Scholar]
  • 5.Shortliffe EHA. rule-based comuter rogram for advising hysicians regarding antimicrobial theray selection. Proceedings of the Annual ACM Conference. ACM; 1974. [Google Scholar]
  • 6.Szolovits P, Pauker SG. Categorical and probabilistic reasoning in medical diagnosis. Artif Intell [Internet] 1978 Aug 1;11(1–2):115–44. https://www.sciencedirect.com/science/article/abs/pii/0004370278900140 (cited 2019 Aug 20) [Google Scholar]
  • 7.Beam AL, Kohane IS. Big data and machine learning in health care. JAMA. 2018;319(13):1317–1318. doi: 10.1001/jama.2017.18391. [DOI] [PubMed] [Google Scholar]
  • 8.Ton J. Cleophas AHZ Machine Learning in Medicine. 2013:1920–30. https://link.springer.com/content/pdf/10.1007%2F978-94-007-5824-7.pdf . [Google Scholar]
  • 9.LeCun Y, Bengio Y, Hinton G. Deep learning. Nature. 2015;521(7553):436–444. doi: 10.1038/nature14539. [DOI] [PubMed] [Google Scholar]
  • 10.Ernande L, Audureau E, Jellis CL, et al. Clinical implications of echocardiographic phenotypes of patients with diabetes mellitus. J Am Coll Cardiol. 2017;70(14):1704–1716. doi: 10.1016/j.jacc.2017.07.792. [DOI] [PubMed] [Google Scholar]
  • 11.Zhao Y, Zeng D, Socinski MA, et al. Reinforcement learning strategies for clinical trials in nonsmall cell lung cancer. Biometrics. 2011;67(4):1422–1433. doi: 10.1111/j.1541-0420.2011.01572.x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Hinton GE, Osindero S, Teh YW. A fast learning algorithm for deep belief nets. Neural Comput. 2006;18(7):1527–1554. doi: 10.1162/neco.2006.18.7.1527. [DOI] [PubMed] [Google Scholar]
  • 13.Chen C. Computer vision in medical imaging. Vol. 2. World Scientific; 2014. [Google Scholar]
  • 14.Puppe F. Systematic introduction to expert systems: Knowledge representations and problem-solving methods. Springer Science & Business Media; 2012. [Google Scholar]
  • 15.Tavel ME. Cardiac auscultation. A glorious past -- but does it have a future? Circulation. 1996;93(6):1250–1253. doi: 10.1161/01.cir.93.6.1250. [DOI] [PubMed] [Google Scholar]
  • 16.Wang A, Bashore TM. Valvular Heart Disease. 2009 [Google Scholar]
  • 17.Thompson WR, Reinisch AJ, Unterberger MJ, et al. Artificial intelligence-assisted auscultation of heart murmurs: validation by virtual clinical trial. Pediatr Cardiol. 2019;40(3):623–629. doi: 10.1007/s00246-018-2036-z. [DOI] [PubMed] [Google Scholar]
  • 18.Dwivedi AK, Imtiaz SA, Rodriguez-Villegas E. Algorithms for automatic analysis and classification of heart sounds-A systematic review. IEEE Access. 2019;7:8316–45. [Google Scholar]
  • 19.Lehner RJ, Rangayyan RM. A three-channel microcomputer system for segmentation and characterization of the phonocardiogram. IEEE Trans Biomed Eng. 1987;34(6):485–489. doi: 10.1109/tbme.1987.326060. [DOI] [PubMed] [Google Scholar]
  • 20.Iwata A, Ishii N, Suzumura N, et al. Algorithm for detecting the first and the second heart sounds by spectral tracking. Med Biol Eng Comput. 1980;18(1):19–26. doi: 10.1007/BF02442475. [DOI] [PubMed] [Google Scholar]
  • 21.Liang H, Lukkarinen S, Hartimo I. Heart sound segmentation algorithm based on heart sound envelogram. 1997;24:105–108. [Google Scholar]
  • 22.Johnson CR, Sethares WA, Klein AG. Software Receiver Design: Build Your Own Digital Communication System in Five Easy Steps. 2011 [Google Scholar]
  • 23.Jiang Z, Choi SA. cardiac sound characteristic waveform method for in-home heart disorder monitoring with electric stethoscope. Expert Syst Appl. 2006;31(2):286–98. [Google Scholar]
  • 24.Sun S. An innovative intelligent system based on automatic diagnostic feature extraction for diagnosing heart diseases. Knowledge-Based Syst. 2015;75:224–238. doi: 10.1016/j.kno-sys.2014.12.001. [DOI] [Google Scholar]
  • 25.Sun S, Jiang Z, Wang H, et al. Automatic moment segmentation and peak detection analysis of heart sound pattern via short-time modified Hilbert transform. Comput Methods Programs Biomed. 2014;114(3):219–230. doi: 10.1016/j.cmpb.2014.02.004. [DOI] [PubMed] [Google Scholar]
  • 26.Gill D, Gavrieli N, Intrator N. Detection and identification of heart sounds using homomorphic envelogram and self-organizing probabilistic model. Comput Cardiol. 2005;32:957–60. [Google Scholar]
  • 27.Springer DB, Tarassenko L, Clifford GD. Logistic Regression-HSMM-Based Heart Sound Segmentation. IEEE Trans Biomed Eng. 2016;63(4):822–832. doi: 10.1109/TBME.2015.2475278. [DOI] [PubMed] [Google Scholar]
  • 28.Schmidt SE, Holst-Hansen C, Graff C, et al. Segmentation of heart sound recordings by a duration-dependent hidden Markov model. Physiol Meas. 2010;31(4):513–529. doi: 10.1088/0967-3334/31/4/004. [DOI] [PubMed] [Google Scholar]
  • 29.Potes C, Parvaneh S, Rahman A, et al. Ensemble of feature-based and deep learning-based classifiers for detection of abnormal heart sounds. Comput Cardiol. 2016;43:621–624. [Google Scholar]
  • 30.Chen TE, Yang SI, Ho LT, et al. S1 and S2 Heart Sound Recognition Using Deep Neural Networks. IEEE Trans Biomed Eng. 2017;64(2):372–380. doi: 10.1109/TBME.2016.2559800. [DOI] [PubMed] [Google Scholar]
  • 31.Zhang W, Han J, Deng S. Heart sound classification based on scaled spectrogram and tensor decomposition. Expert Syst Appl. 2017;84:220–231. [Google Scholar]
  • 32.Khalid S, Khalil T, Nasreen SA. survey of feature selection and feature extraction techniques in machine learning. Proc 2014 Sci Inf Conf SAI; 2014. pp. 372–378. [Google Scholar]
  • 33.Zhang W, Guo X, Yuan Z, et al. Heart sound classification and recognition based on eemd and correlation dimension. J Mech Med Biol. 2014;14(04):1450046. [Google Scholar]
  • 34.Safara F, Doraisamy S, Azman A, et al. Multi-level basis selection of wavelet packet decomposition tree for heart sound classification. Comput Biol Med. 2013;43(10):1407–1414. doi: 10.1016/j.compbiomed.2013.06.016. [DOI] [PubMed] [Google Scholar]
  • 35.Kwak C, Kwon OW. Cardiac disorder classification by heart sound signals using murmur likelihood and hidden Markov model state likelihood. IET Signal Process. 2012;6(4):326. [Google Scholar]
  • 36.Kumar D, Carvalho P, Antunes M, et al. Heart murmur classification with feature selection. Conf Proc IEEE Eng Med Biol Soc.; 2010. pp. 4566–4569. [DOI] [PubMed] [Google Scholar]
  • 37.Uğuz H. A biomedical system based on artificial neural network and principal component analysis for diagnosis of the heart valve diseases. J Med Syst. 2012;36(1):61–72. doi: 10.1007/s10916-010-9446-7. [DOI] [PubMed] [Google Scholar]
  • 38.Olmez T, Zumray D. Classification of heart sounds using an artificial neural network. Pattern Recognit Lett. 2003;24:617–629. [Google Scholar]
  • 39.Fahad HM, Ghani Khan MU, Saba T, et al. Microscopic abnormality classification of cardiac murmurs using ANFIS and HMM. Microsc Res Tech. 2018;81(5):449–457. doi: 10.1002/jemt.22998. [DOI] [PubMed] [Google Scholar]
  • 40.Falk V, Baumgartner H, Bax JJ, et al. ESC/EACTS Guidelines for the management of valvular heart disease. Eur J Cardio-thoracic Surg. 2017;52:616–664. doi: 10.1093/ejcts/ezx324. [DOI] [PubMed] [Google Scholar]
  • 41.Noble JA, Boukerroui D. Ultrasound image segmentation: a survey. IEEE Trans Med Imaging. 2006;25(8):987–1010. doi: 10.1109/tmi.2006.877092. [DOI] [PubMed] [Google Scholar]
  • 42.Mikić I, Krucinski S, Thomas JD. Segmentation and tracking in echocardiographic sequences: active contours guided by optical flow estimates. IEEE Trans Med Imaging. 1998;17(2):274–284. doi: 10.1109/42.700739. [DOI] [PubMed] [Google Scholar]
  • 43.Martin S, Daanen V, Troccaz J, et al. Tracking of the mitral valve leaflet in echocardiography images. 3rd IEEE Int Symp Biomed Imaging Nano to Macro; 2006. pp. 181–184. [Google Scholar]
  • 44.Schneider RJ, Perrin DP, Vasilyev NV, et al. Mitral annulus segmentation from 3D ultrasound using graph cuts. IEEE Trans Med Imaging. 2010;29(9):1676–1687. doi: 10.1109/TMI.2010.2050595. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 45.Zhou X, Yang C, Yu W. Automatic mitral leaflet tracking in echocardiography by outlier detection in the low-rank representation. Proc IEEE Comput Soc Conf Comput Vis Pattern Recognit; 2012. pp. 972–979. [Google Scholar]
  • 46.Liu X, Cheung Y, Ming Y, et al. Automatic mitral valve leaflet tracking in Echocardiography via constrained outlier pursuit and region-scalable active contours. Neurocomputing. 2014;144:47–57. doi: 10.1016/j.neucom.2014.02.063. [DOI] [Google Scholar]
  • 47.Siefert AW, Icenogle DA, Rabbah JPM, et al. Accuracy of a mitral valve segmentation method using J-splines for real-time 3D echocardiography data. Ann Biomed Eng. 2013;41(6):1258–1268. doi: 10.1007/s10439-013-0784-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 48.Pouch AM, Wang H, Takabe M, et al. Fully automatic segmentation of the mitral leaflets in 3D transesophageal echocardiographic images using multi-atlas joint label fusion and deformable medial modeling. Med Image Anal. 2014;18(1):118–129. doi: 10.1016/j.media.2013.10.001. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 49.Ionasec RI, Voigt I, Georgescu B, et al. Patient-specific modeling and quantification of the aortic and mitral valves from 4-D cardiac CT and TEE. IEEE Trans Med Imaging. 2010;29(9):1636–1651. doi: 10.1109/TMI.2010.2048756. [DOI] [PubMed] [Google Scholar]
  • 50.Costa E, Martins N, Sultan MS, et al. Mitral valve leaflets segmentation in echocardiography using convolutional neural networks. 6th IEEE Port Meet Bioeng ENBENG 2019 Proc; 2019. pp. 1–4. [Google Scholar]
  • 51.Shang Y, Yang X, Zhu L, et al. Region competition based active contour for medical object extraction. Comput Med Imaging Graph. 2008;32(2):109–117. doi: 10.1016/j.compmed-imag.2007.10.004. [DOI] [PubMed] [Google Scholar]
  • 52.Sultan MS, Martins N, Costa E, et al. Virtual M-Mode for Echo-cardiography: A New Approach for the Segmentation of the Anterior Mitral Leaflet. IEEE J Biomed Health Inform. 2019;23(1):305–313. doi: 10.1109/JBHI.2018.2799738. [DOI] [PubMed] [Google Scholar]
  • 53.Thavendiranathan P, Phelan D, Thomas JD, et al. Quantitative assessment of mitral regurgitation: validation of new methods. J Am Coll Cardiol. 2012;60(16):1470–1483. doi: 10.1016/j.jacc.2012.05.048. [DOI] [PubMed] [Google Scholar]
  • 54.de Agustín JA, Marcos-Alberca P, Fernandez-Golfin C, et al. Direct measurement of proximal isovelocity surface area by single-beat three-dimensional color Doppler echocardiography in mitral regurgitation: a validation study. J Am Soc Echocardiogr. 2012;25(8):815–823. doi: 10.1016/j.echo.2012.05.021. [DOI] [PubMed] [Google Scholar]
  • 55.Grady L, Datta S, Kutter O, et al. Regurgitation Quantification Using 3D PISA in Volume Echocardiograhy. In: Fichtinger G, editor. pMedical Image Comuting and Comuter-AssisteIntervention -- MICCAI 2011. Berlin, Heidelberg: Sringer Berlin Heidelberg; 2011. pp. 512–519. [DOI] [PubMed] [Google Scholar]
  • 56.Cobey FC, McInnis JA, Gelfand BJ, et al. A method for automating 3-dimensional proximal isovelocity surface area measurement. J Cardiothorac Vasc Anesth. 2012;26(3):507–511. doi: 10.1053/j.jvca.2011.12.018. [DOI] [PubMed] [Google Scholar]
  • 57.Wang Y, Georgescu B, Datta S, et al. Automatic cardiac flow quantification on 3D volume color Doppler data. Proc - Int Symp Biomed Imaging. 2011;C(Lv):1688–1691. [Google Scholar]
  • 58.Mediratta A, Addetia K, Medvedofsky D, et al. 3D echocardiographic analysis of aortic annulus for transcatheter aortic valve replacement using novel aortic valve quantification software: Comparison with computed tomography. Echocardiography. 2017;34(5):690–699. doi: 10.1111/echo.13483. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 59.Aquila I, Fernández-Golfín C, Rincon LM, et al. Fully automated software for mitral annulus evaluation in chronic mitral regurgitation by 3-dimensional transesophageal echocardiography. Medicine (Baltimore) 2016;95(49):e5387. doi: 10.1097/MD.0000000000005387. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 60.Kagiyama N, Toki M, Hara M, et al. Efficacy and Accuracy of Novel Automated Mitral Valve Quantification: Three-Dimensional Transesophageal Echocardiographic Study. Echocardiography. 2016;33(5):756–763. doi: 10.1111/echo.13135. [DOI] [PubMed] [Google Scholar]
  • 61.Calleja A, Thavendiranathan P, Ionasec RI, et al. Automated quantitative 3-dimensional modeling of the aortic valve and root by 3-dimensional transesophageal echocardiography in normals, aortic regurgitation, and aortic stenosis: comparison to computed tomography in normals and clinical implications. Circ Cardiovasc Imaging. 2013;6(1):99–108. doi: 10.1161/CIRCIMAGING.112.976993. [DOI] [PubMed] [Google Scholar]
  • 62.de Agustin JA, Viliani D, Vieira C, et al. Proximal isovelocity surface area by single-beat three-dimensional color Doppler echo-cardiography applied for tricuspid regurgitation quantification. J Am Soc Echocardiogr. 2013;26(9):1063–1072. doi: 10.1016/j.echo.2013.06.006. [DOI] [PubMed] [Google Scholar]
  • 63.Choi J, Hong GR, Kim M, et al. Automatic quantification of aortic regurgitation using 3D full volume color doppler echocardiography: a validation study with cardiac magnetic resonance imaging. Int J Cardiovasc Imaging. 2015;31(7):1379–1389. doi: 10.1007/s10554-015-0707-x. [DOI] [PubMed] [Google Scholar]
  • 64.Thavendiranathan P, Liu S, Datta S, et al. Quantification of chronic functional mitral regurgitation by automated 3-dimensional peak and integrated proximal isovelocity surface area and stroke volume techniques using real-time 3-dimensional volume color doppler echocardiography: In vitro and clini. Circ Cardiovasc Imaging. 2013;6(1):125–133. doi: 10.1161/CIRCIMAGING.112.980383. [DOI] [PubMed] [Google Scholar]
  • 65.Ionasec RI, Georgescu B, Gassner E, et al. Dynamic Model-Driven Quantitative and Visual Evaluation of the Aortic Valve from 4D CT. In: Metaxas D, editor. Medical Image Comuting and Comuter-Assisted Intervention. MICCAI 2008. Berlin, Heidelberg: Sringer Berlin Heidelberg; 2008. pp. 686–694. [DOI] [PubMed] [Google Scholar]
  • 66.Swee JKY, Grbić S. Advanced transcatheter aortic valve implantation (TAVI) planning from CT with ShapeForest. Lect Notes Comput Sci. (including Subser Lect Notes Artif Intell Lect Notes Bioinformatics) 2014;8674 LNCS(Part 2):17–24. doi: 10.1007/978-3-319-10470-6_3. [DOI] [PubMed] [Google Scholar]
  • 67.Waechter I, Kneser R, Korosoglou G, et al. Patient Secific Models for Planning and Guidance of Minimally Invasive Aortic Valve Imlantation. In: Jiang T, Navab N, Pluim JPW, Viergever MA, editors. Medical Image Comuting and Comuter-Assisted Intervention -- MICCAI 2010. Sringer Berlin Heidelberg; 2010. pp. 526–533. [DOI] [PubMed] [Google Scholar]
  • 68.Grbic S, Ionasec R, Vitanovski D, et al. Complete valvular heart apparatus model from 4D cardiac CT. Med Image Anal. 2012;16(5):1003–1014. doi: 10.1016/j.media.2012.02.003. [DOI] [PubMed] [Google Scholar]
  • 69.Grbic S, Ionasec R, Mansi T, et al. Advanced intervention planning for Transcatheter Aortic Valve Implantations (TAVI) from CT using volumetric models. Proc Int Symp Biomed Imaging. 2013:1424–1427. [Google Scholar]
  • 70.Koos R, Mahnken AH, Dohmen G, et al. Association of aortic valve calcification severity with the degree of aortic regurgitation after transcatheter aortic valve implantation. Int J Cardiol. 2011;150(2):142–145. doi: 10.1016/j.ijcard.2010.03.004. [DOI] [PubMed] [Google Scholar]
  • 71.Samim M, Stella PR, Agostoni P, et al. Automated 3D analysis of pre-procedural MDCT to predict annulus plane angulation and C-arm positioning: benefit on procedural outcome in patients referred for TAVR. JACC Cardiovasc Imaging. 2013;6(2):238–248. doi: 10.1016/j.jcmg.2012.12.004. [DOI] [PubMed] [Google Scholar]
  • 72.Zheng Y, John M, Liao R, et al. Automatic aorta segmentation and valve landmark detection in C-arm CT for transcatheter aortic valve implantation. IEEE Trans Med Imaging. 2012;31(12):2307–2321. doi: 10.1109/TMI.2012.2216541. [DOI] [PubMed] [Google Scholar]
  • 73.Elattar M, Wiegerinck E, van Kesteren F, et al. Automatic aortic root landmark detection in CTA images for preprocedural planning of transcatheter aortic valve implantation. Int J Cardiovasc Imaging. 2016;32(3):501–511. doi: 10.1007/s10554-015-0793-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 74.Elattar MA, Wiegerinck EM, Planken RN, et al. Automatic segmentation of the aortic root in CT angiography of candidate patients for transcatheter aortic valve implantation. Med Biol Eng Comput. 2014;52(7):611–618. doi: 10.1007/s11517-014-1165-7. [DOI] [PubMed] [Google Scholar]
  • 75.Lalys F, Esneault S, Castro M, et al. Automatic aortic root segmentation and anatomical landmarks detection for TAVI procedure planning. Minim Invasive Ther Allied Technol. 2019;28(3):157–164. doi: 10.1080/13645706.2018.1488734. [DOI] [PubMed] [Google Scholar]
  • 76.Al WA, Jung HoY, Yun IlD, et al. Automatic aortic valve landmark localization in coronary CT angiography using colonial walk. PLoS One. 2018;13(7):e0200317. doi: 10.1371/journal.pone.0200317. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 77.Watanabe Y, Morice MC, Bouvier E, et al. Automated 3-dimensional aortic annular assessment by multidetector computed tomography in transcatheter aortic valve implantation. JACC Cardiovasc Interv. 2013;6(9):955–964. doi: 10.1016/j.jcin.2013.05.008. [DOI] [PubMed] [Google Scholar]
  • 78.Delgado V, Ng ACT, Schuijf JD, et al. Automated assessment of the aortic root dimensions with multidetector row computed tomography. Ann Thorac Surg. 2011;91(3):716–723. doi: 10.1016/j.athoracsur.2010.09.060. [DOI] [PubMed] [Google Scholar]
  • 79.Dankerl P, Hammon M, Seuss H, et al. Computer-aided evaluation of low-dose and low-contrast agent third-generation dualsource CT angiography prior to transcatheter aortic valve implantation (TAVI) Int J Comput Assist Radiol Surg. 2017;12(5):795–802. doi: 10.1007/s11548-016-1470-8. [DOI] [PubMed] [Google Scholar]
  • 80.Guez D, Boroumand G, Ruggiero NJ, et al. Automated and Manual Measurements of the Aortic Annulus with ECG-Gated Cardiac CT Angiography Prior to Transcatheter Aortic Valve Replacement: Comparison with 3D-Transesophageal Echocardiography. Acad Radiol. 2017;24(5):587–593. doi: 10.1016/j.acra.2016.12.008. [DOI] [PubMed] [Google Scholar]
  • 81.Baeßler B, Mauri V, Bunck AC, et al. Software-automated multidetector computed tomography-based prosthesis-sizing in transcatheter aortic valve replacement: Inter-vendor comparison and relation to patient outcome. Int J Cardiol. 2018;272:267–272. doi: 10.1016/j.ijcard.2018.07.008. [DOI] [PubMed] [Google Scholar]
  • 82.Pennell DJ. Cardiovascular magnetic resonance. Circulation. 2010;121(5):692–705. doi: 10.1161/CIRCULATIONAHA.108.811547. [DOI] [PubMed] [Google Scholar]
  • 83.Bernard O, Cervenansky F, Lalande A, et al. Deep learning techniques for automatic MRI cardiac multi-structures segmentation and diagnosis: is the problem solved? IEEE Trans Med Imaging. 2018;37(11):2514–2525. doi: 10.1109/TMI.2018.2837502. [DOI] [PubMed] [Google Scholar]
  • 84.Petitjean C, Dacher JN. A review of segmentation methods in short axis cardiac MR images. Med Image Anal. 2011;15(2):169–184. doi: 10.1016/j.media.2010.12.004. [DOI] [PubMed] [Google Scholar]
  • 85.Backhaus SJ, Staab W, Steinmetz M, et al. Fully automated quantification of biventricular volumes and function in cardiovascular magnetic resonance: applicability to clinical routine settings. J Cardiovasc Magn Reson. 2019;21(1):24. doi: 10.1186/s12968-019-0532-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 86.Fries JA, Varma P, Chen VS, et al. Weakly supervised classification of rare aortic valve malformations using unlabeled cardiac MRI sequences. bioRxiv. 2018;2019:1–25. doi: 10.1038/s41467-019-11012-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 87.Nashef SAM, Roques F, Sharples LD, et al. EuroSCORE II. Eur J Cardiothorac Surg. 2012;41(4):734–44. doi: 10.1093/ejcts/ezs043. discussion 744. [DOI] [PubMed] [Google Scholar]
  • 88.Nilsson J, Ohlsson M, Thulin L, et al. Risk factor identification and mortality prediction in cardiac surgery using artificial neural networks. J Thorac Cardiovasc Surg. 2006;132(1):12–19. doi: 10.1016/j.jtcvs.2005.12.055. [DOI] [PubMed] [Google Scholar]
  • 89.Celi LA, Galvin S, Davidzon G, et al. A database-driven decision support system: customized mortality prediction. J Pers Med. 2012;2(4):138–148. doi: 10.3390/jpm2040138. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 90.Allyn J, Allou N, Augustin P, et al. A Comparison of a Machine Learning Model with EuroSCORE II in Predicting Mortality after Elective Cardiac Surgery: A Decision Curve Analysis. PLoS One. 2017;12(1):e0169772. doi: 10.1371/journal.pone.0169772. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 91.Hernandez-Suarez DF, Kim Y, Villablanca P, et al. Machine Learning Prediction Models for In-Hospital Mortality After Transcatheter Aortic Valve Replacement. JACC Cardiovasc Interv. 2019;12(14):1328–1338. doi: 10.1016/j.jcin.2019.06.013. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 92.LaFaro RJ, Pothula S, Kubal KP, et al. Neural Network Prediction of ICU Length of Stay Following Cardiac Surgery Based on Pre-Incision Variables. PLoS One. 2015;10(12):e0145395. doi: 10.1371/journal.pone.0145395. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 93.Thottakkara P, Ozrazgat-Baslanti T, Hupf BB, et al. Application of Machine Learning Techniques to High-Dimensional Clinical Data to Forecast Postoperative Complications. PLoS One. 2016;11(5):e0155705. doi: 10.1371/journal.pone.0155705. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 94.Nilsson J, Ohlsson M, Höglund P, et al. The International Heart Transplant Survival Algorithm (IHTSA): a new model to improve organ sharing and survival. PLoS One. 2015;10(3):e0118644. doi: 10.1371/journal.pone.0118644. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 95.Moustris GP, Hiridis SC, Deliparaschos KM, et al. Evolution of autonomous and semi-autonomous robotic surgical systems: a review of the literature. Int J Med Robot. 2011;7(4):375–392. doi: 10.1002/rcs.408. [DOI] [PubMed] [Google Scholar]
  • 96.Grespan L, Fiorini P, Colucci G. The Route to Patient Safety in Robotic Surgery [Internet] Cham: Springer International Publishing; 2019. Looking Ahead: The Future of Robotic Surgery; pp. 157–162. [DOI] [Google Scholar]
  • 97.Panesar S, Cagle Y, Chander D, et al. Artificial Intelligence and the Future of Surgical Robotics. Ann Surg. 2019;270(2):223–226. doi: 10.1097/SLA.0000000000003262. [DOI] [PubMed] [Google Scholar]
  • 98.Garcia-Ruiz A, Gagner M, Miller JH, et al. Manual vs robotically assisted laparoscopic surgery in the performance of basic manipulation and suturing tasks. Arch Surg. 1998;133(9):957–961. doi: 10.1001/archsurg.133.9.957. [DOI] [PubMed] [Google Scholar]
  • 99.Mayer H, Gomez F, Wierstra D, et al. system for robotic heart surgery that learns to tie knots using recurrent neural networks. Adv Robot. 2008;13(14):1521–1537. [Google Scholar]
  • 100.Fagogenis G, Mencattelli M, Machaidze Z, et al. Autonomous Robotic Intracardiac Catheter Navigation Using Haptic Vision. Sci Robot. 2019;4(29) doi: 10.1126/scirobotics.aaw1977. [DOI] [PMC free article] [PubMed] [Google Scholar]

Articles from Cardiology Journal are provided here courtesy of Via Medica sp. z o.o. sp. k.

RESOURCES