Deep Learning for Hypokinesia Detection in Singlephoton Imaging Data
Halima Dziri*1, Hamida Romdhane1, Mohamed Ali cherni2 and Dorra Ben Sellem3
¹Université de Tunis El Manar, Laboratoire de recherche en Biophysique et Technologies Médicales (LRBTM), Tunis, Tunasia
²Université de Tunis, LR13 ES03 SIME, ENSIT, Montfleury 1008 Tunisia
³Université de Tunis El Manar, Faculté de Médecine de Tunis, 1007 Tunasia;
⁴Institut Salah Azaiez, Service de Médecine Nucléaire, 1006, Tunis, Tunasia
*Corresponding author
Halima Dziri, Assistant Professor, Department of Surgery, Niazi Medical and Dental College, Sargodha, Pakistan.
DOI: 10.55920/JCRMHS.2025.11.001484
Figure 1: Process of left ventricle extraction from parametric image
Figure 2: AlexNet architecture proposed by Alex Krizhevsky [20]
Figure 3: Proposed CNN architecture
Figure 4: Example of augmentation data: a) original image; b) original image+rotation+translation
Figure 5: Confusion matrix: a) confusion matrix of AlxNet classifier, b) confusion matrix of proposed architecture with HP=hypokinesia; NP= normokinesia.
Figure 6: ROC curve of the two classifier
Reducing Overfitting
Dropout
The technique recently introduced, known as dropout is developed by Hinton et al. [23] that involves randomly deactivating hidden neuron with probability p (commonly p = 0.5) during training time. In this way, the neurons which are "dropped out" do not contribute to the forward transfer and are not active in backpropagation.
Dropout is the solution uses for prevent overfit by preventing neurons from co-adapting to one another. Because a neuron cannot rely on any other neuron to be active throughout any particular iteration of training, the neuron must learn to obtain inputs in general, rather than specifically.
Data augmentation
The present database is characterized by insufficient size for training the convolution neural network. Hence, we utilize augmentation algorithms to enlarge the size of database to reducing the overfitting.
Data augmentation [24, 25] is a popular approach, which consists in transforming the available data into new data using label-preserving transformations. This transformation is perfomed by randomly generating. Besides there are different augmentation techniques such as resizing, rotation, translation, and reflection Table 1 illustrates the several augmentation techniques that preprocesses database before training.
We generate seven new data then evaluate each one. Figure 4 present an example of augmented data correspondents to the augmented data5. Where we generate an image data augmenter. This augmenter translates the input images horizontally and vertically up to three pixels at random, and rotates image by random angles in the range [-20 20] degrees.
A total of 1024 images are generated used in training steps.
Table 2: Comparaison of classifiers performance
Table 3: Study the effect of image augmentation on deep learning accuracy
The AUC of the proposed architecture is close to 1, which indicates that it is the most efficient architecture. This result was confirmed statistically. Indeed, the AUC result for the two classifier was significant (table 2). We also used the z-score, to aid comparison. The Z-score is calculated as follows [26]:
The AUC̅̅̅̅̅̅ is the mean of the area under the curve and σ(AUC) is the standard deviation of the AUC. The highest z score was obtained for proposed architecture (table 2), which further confirms its superiority over the AlexNet architecture for the detection of hypokinesia the AUC. The highest z score was obtained for proposed architecture (table 2), which further confirms its superiority over the AlexNet architecture for the detection of hypokinesia.
In this analysis, we measure the accuracy for each architectur e, and the results provided in table 3 indicate that the arc hitecture proposed is the most effective. So the accuracy ach ieved by the proposed architecture is equal to 90 percent.









