2023, Articolo in rivista, ENG
Daniel Riccio and Nadia Brancati and Giovanna Sannino and Laura Verde and Maria Frucci
Deep Learning based heart sound classification is of significant interest in reducing the burden of manual auscultation through the automated detection of signals, including abnormal heartbeats. This work presents a method for classifying phonocardiogram (PCG) signals as normal or abnormal by applying a deep Convolutional Neural Network (CNN) after transforming the signals into 2D color images. In particular, a new methodology based on fractal theory, which exploits Partitioned Iterated Function Systems (PIFS) to generate 2D color images from 1D signals is presented. PIFS have been extensively investigated in the context of image coding and indexing on account of their ability to interpolate and identify self-similar features in an image. Our classification approach has shown a high potential in terms of noise robustness and does not require any pre-processing steps or an initial segmentation of the signal, as instead happens in most of the approaches proposed in the literature. In this preliminary work, we have carried out several experiments on the database released for the 2016 Physionet Challenge, both in terms of different classification networks and different inputs to the networks, thus also evaluating the data quality. Among all experiments, we have obtained the best result of 0.85 in terms of modified Accuracy (MAcc).
2023, Articolo in rivista, ENG
Mosaiyebzadeh, Fatemeh; Pouriyeh, Seyedamin; Parizi, Reza M.; Sheng, Quan Z.; Han, Meng; Zhao, Liang; Sannino, Giovanna; Ranieri, Caetano Mazzoni; Ueyama, Jo; Batista, Daniel Macedo
Advancements in wearable medical devices using the IoT technology are shaping the modern healthcare system. With the emergence of the Internet of Healthcare Things (IoHT), efficient healthcare services can be provided to patients. Healthcare professionals have effectively used AI-based models to analyze the data collected from IoHT devices to treat various diseases. Data must be processed and analyzed while avoiding privacy breaches, in compliance with legal rules and regulations, such as the HIPAA and GDPR. Federated learning (FL) is a machine learning-based approach allowing multiple entities to train an ML model collaboratively without sharing their data. It is particularly beneficial in healthcare, where data privacy and security are substantial concerns. Even though FL addresses some privacy concerns, there is still no formal proof of privacy guarantees for IoHT data. Privacy-enhancing technologies (PETs) are tools and techniques designed to enhance the privacy and security of online communications and data sharing. PETs provide a range of features that help protect users' personal information and sensitive data from unauthorized access and tracking. This paper comprehensively reviews PETs concerning FL in the IoHT scenario and identifies several key challenges for future research.
2022, Contributo in volume, ENG
Verde L.; Sannino G.
This chapter presents an overview of the main Artificial Intelligence models used for time series data analysis, highlighting the main characteristics of each. The aim is to provide researchers with an panoramic that can guide them in choosing the most suitable technique for their studies.
2022, Articolo in rivista, ENG
De Falco I.; De Pietro G.; Sannino G.
The classification of images is of high importance in medicine. In this sense, Deep learning methodologies show excellent performance with regard to accuracy. The drawback of these methodologies is the fact that they are black boxes, so no explanation is given to users on the reasons underlying their choices. In the medical domain, this lack of transparency and information, typical of black box models, brings practitioners to raise concerns, and the result is a resistance to the use of deep learning tools. In order to overcome this problem, a different Machine Learning approach to image classification is used here that is based on interpretability concepts thanks to the use of an evolutionary algorithm. It relies on the application of two steps in succession. The first receives a set of images in the inut and performs image filtering on them so that a numerical data set is generated. The second is a classifier, the kernel of which is an evolutionary algorithm. This latter, at the same time, classifies and automatically extracts explicit knowledge as a set of IF-THEN rules. This method is investigated with respect to a data set of MRI brain imagery referring to Alzheimer's disease. Namely, a two-class data set (non-demented and moderate demented) and a three-class data set (non-demented, mild demented, and moderate demented) are extracted. The methodology shows good results in terms of accuracy (100% for the best run over the two-class problem and 91.49% for the best run over the three-class one), F_score (1.0000 and 0.9149, respectively), and Matthews Correlation Coefficient (1.0000 and 0.8763, respectively). To ascertain the quality of these results, they are contrasted against those from a wide set of well-known classifiers. The outcome of this comparison is that, in both problems, the methodology achieves the best results in terms of accuracy and F_score, whereas, for the Matthews Correlation Coefficient, it has the best result over the two-class problem and the second over the three-class one.
DOI: 10.3390/s22113966
2022, Editoriale in rivista, ENG
Giovanna Sannino, Nadia Brancati, Alfred M. Bruckstein, Maria Frucci, Daniel Riccio
Biomedical signal processing and control (Print) 722022, Articolo in rivista, ENG
De Falco I.; De Pietro G.; Sannino G.
In medical practice, all decisions, as for example the diagnosis based on the classification of images, must be made reliably and effectively. The possibility of having automatic tools helping doctors in performing these important decisions is highly welcome. Artificial Intelligence techniques, and in particular Deep Learning methods, have proven very effective on these tasks, with excellent performance in terms of classification accuracy. The problem with such methods is that they represent black boxes, so they do not provide users with an explanation of the reasons for their decisions. Confidence from medical experts in clinical decisions can increase if they receive from Artificial Intelligence tools interpretable output under the form of, e.g., explanations in natural language or visualized information. This way, the system outcome can be critically assessed by them, and they can evaluate the trustworthiness of the results. In this paper, we propose a new general-purpose method that relies on interpretability ideas. The approach is based on two successive steps, the former being a filtering scheme typically used in Content-Based Image Retrieval, whereas the latter is an evolutionary algorithm able to classify and, at the same time, automatically extract explicit knowledge under the form of a set of IF-THEN rules. This approach is tested on a set of chest X-ray images aiming at assessing the presence of COVID-19.
2022, Articolo in rivista, ENG
L. Verde, N. Brancati, G. De Pietro, M. Frucci, G. Sannino
Edge Analytics and Artificial Intelligence are important features of the current smart connected living community. In a society where people, homes, cities, and workplaces are simultaneously connected through various devices, primarily through mobile devices, a considerable amount of data is exchanged, and the processing and storage of these data are laborious and difficult tasks. Edge Analytics allows the collection and analysis of such data on mobile devices, such as smartphones and tablets, without involving any cloud-centred architecture that cannot guarantee real-time responsiveness. Meanwhile, Artificial Intelligence techniques can constitute a valid instrument to process data, limiting the computation time, and optimising decisional processes and predictions in several sectors, such as healthcare. Within this field, in this article, an approach able to evaluate the voice quality condition is proposed. A fully automatic algorithm, based on Deep Learning, classifies a voice as healthy or pathological by analysing spectrogram images extracted by means of the recording of vowel /a/, in compliance with the traditional medical protocol. A light Convolutional Neural Network is embedded in a mobile health application in order to provide an instrument capable of assessing voice disorders in a fast, easy, and portable way. Thus, a straightforward mobile device becomes a screening tool useful for the early diagnosis, monitoring, and treatment of voice disorders. The proposed approach has been tested on a broad set of voice samples, not limited to the most common voice diseases but including all the pathologies present in three different databases achieving F1-scores, over the testing set, equal to 80%, 90%, and 73%. Although the proposed network consists of a reduced number of layers, the results are very competitive compared to those of other "cutting edge" approaches constructed using more complex neural networks, and compared to the classic deep neural networks, for example, VGG-16 and ResNet-50.
DOI: 10.1145/3433993
2021, Contributo in atti di convegno, ENG
Giovanna Sannino; Giuseppe De Pietro; Ivanoe De Falco
Cardiovascular diseases cause the death of almost 18 million people each year. Heart failure takes place when the heart does not pump a sufficient amount of blood to the body and is one of the most common causes of death. Healthcare technologies and Artificial Intelligence algorithms constitute promising solutions useful not only for the monitoring of people's health but also in order to tell which subjects are more prone to this problem, which is information of paramount relevance to save their lives. The goal of this paper is to understand the predictability of mortality of subjects suffering from left ventricular systolic dysfunction who previously experienced heart failures. To perform this important study, a publicly-available data set is considered that contains thirteen pieces of clinical, body, and lifestyle information about 299 subjects. In tackling this data set, not only do we wish to perform classification with reference to subjects' survival/death, but we also wish to automatically extract explainable knowledge about the reasons for the classification proposed. To this aim, we use DEREx, an Artificial Intelligence-based tool that relies on Evolutionary Algorithms and provides users with an easy-to-understand set of IF-THEN rules containing data set parameters. In this way, it performs the selection of the parameters that are the most relevant for the purpose of classification. We have run our experiments following a sound protocol established in the scientific literature for this data set. Our findings show that, apart from automatically obtaining easily interpretable knowledge, DEREx achieves better results in terms of widely-used quality indices as Matthews Correlation Coefficient, accuracy, and F score.
2021, Editoriale in rivista, ENG
Celesti A.; De Falco I.; Pecchia L.; Sannino G.
IEEE Journal of Biomedical and Health Informatics 25, pp. 4240–42422021, Articolo in rivista, ENG
Verde, Laura; De Pietro, Giuseppe; Ghoneim, Ahmed; Alrashoud, Mubarak; Al-Mutib, Khaled N.; Sannino, Giovanna
The Covid-19 pandemic represents one of the greatest global health emergencies of the last few decades with indelible consequences for all societies throughout the world. The cost in terms of human lives lost is devastating on account of the high contagiousness and mortality rate of the virus. Millions of people have been infected, frequently requiring continuous assistance and monitoring. Smart healthcare technologies and Artificial Intelligence algorithms constitute promising solutions useful not only for the monitoring of patient care but also in order to support the early diagnosis, prevention and evaluation of Covid-19 in a faster and more accurate way. On the other hand, the necessity to realise reliable and precise smart healthcare solutions, able to acquire and process voice signals by means of appropriate Internet of Things devices in real-time, requires the identification of algorithms able to discriminate accurately between pathological and healthy subjects. In this paper, we explore and compare the performance of the main machine learning techniques in terms of their ability to correctly detect Covid-19 disorders through voice analysis. Several studies report, in fact, significant effects of this virus on voice production due to the considerable impairment of the respiratory apparatus. Vocal folds oscillations that are more asynchronous, asymmetrical and restricted are observed during phonation in Covid-19 patients. Voice sounds selected by the Coswara database, an available crowd-sourced database, have been e analysed and processed to evaluate the capacity of the main ML techniques to distinguish between healthy and pathological voices. All the analyses have been evaluated in terms of accuracy, sensitivity, specificity, F1-score and Receiver Operating Characteristic area. These show the reliability of the Support Vector Machine algorithm to detect the Covid-19 infections, achieving an accuracy equal to about 97%.
2021, Contributo in volume, ENG
Sannino, Giovanna; De Falco, Ivanoe
The absence of the rapid eye movement (REM) phase during sleep can have negative consequences, as, e.g., anxiety, increase in appetite, irritability, while, on the other hand, it can help in improving some kinds of depression. The goal of the research described in this chapter consists in the identification of the different sleep phases a subject is experiencing by using heart rate variability (HRV) values. These are computed starting from the signals gathered from electrocardiogram (ECG) sensors placed on the subject. To this aim, the publicly available Sleep Heart Health Study (SHHS) data set is taken into account, which contains both types of information. Several machine learning classification algorithms are tested on this data set, and their performance is compared in terms of F1-score value, as SHHS is highly unbalanced. Once the most suitable classification algorithm is found, it can be firstly trained offline on the problem and then used online in an IoT-based fully automated e-health system. In this latter, sensors gather, in real time, ECG signals from a sleeping subject, send them to a device where data is processed, HRV values are computed, sleep phase identification takes place, and medical personnel, close or not to the subject, are immediately informed of the subject's sleeping phases.
2021, Editoriale in rivista, ENG
Celesti, Antonio; De Falco, Ivanoe; Galletta, Antonino; Sannino, Giovanna
Computers (Basel) 10 (8)2021, Articolo in rivista, ENG
Verde, Laura; De Pietro, Giuseppe; Sannino, Giovanna
Healthcare sensors represent a valid and non-invasive instrument to capture and analyse physiological data. Several vital signals, such as voice signals, can be acquired anytime and anywhere, achieved with the least possible discomfort to the patient thanks to the development of increasingly advanced devices. The integration of sensors with artificial intelligence techniques contributes to the realization of faster and easier solutions aimed at improving early diagnosis, personalized treatment, remote patient monitoring and better decision making, all tasks vital in a critical situation such as that of the COVID-19 pandemic. This paper presents a study about the possibility to support the early and non-invasive detection of COVID-19 through the analysis of voice signals by means of the main machine learning algorithms. If demonstrated, this detection capacity could be embedded in a powerful mobile screening application. To perform this important study, the Coswara dataset is considered. The aim of this investigation is not only to evaluate which machine learning technique best distinguishes a healthy voice from a pathological one, but also to identify which vowel sound is most seriously affected by COVID-19 and is, therefore, most reliable in detecting the pathology. The results show that Random Forest is the technique that classifies most accurately healthy and pathological voices. Moreover, the evaluation of the vowel /e/ allows the detection of the effects of COVID-19 on voice quality with a better accuracy than the other vowels.
2021, Articolo in rivista, ENG
Muhammad, Khan; Mustaqeem, null; Ullah, Amin; Imran, Ali Shariq; Sajjad, Muhammad; Kiran, Mustafa Servet; Sannino, Giovanna; de Albuquerque, Victor Hugo C.
Human action recognition in videos is an active area of research in computer vision and pattern recognition. Nowadays, artificial intelligence (AI) based systems are needed for human-behavior assessment and security purposes. The existing action recognition techniques are mainly using pre-trained weights of different AI architectures for the visual representation of video frames in the training stage, which affect the features' discrepancy determination, such as the distinction between the visual and temporal signs. To address this issue, we propose a bi-directional long short-term memory (BiLSTM) based attention mechanism with a dilated convolutional neural network (DCNN) that selectively focuses on effective features in the input frame to recognize the different human actions in the videos. In this diverse network, we use the DCNN layers to extract the salient discriminative features by using the residual blocks to upgrade the features that keep more information than a shallow layer. Furthermore, we feed these features into a BiLSTM to learn the long-term dependencies, which is followed by the attention mechanism to boost the performance and extract the additional high-level selective action related patterns and cues. We further use the center loss with Softmax to improve the loss function that achieves a higher performance in the video-based action classification. The proposed system is evaluated on three benchmarks, i.e., UCF11, UCF sports, and J-HMDB datasets for which it achieved a recognition rate of 98.3%, 99.1%, and 80.2%, respectively, showing 1%-3% improvement compared to the state-of-the-art (SOTA) methods.
2020, Contributo in atti di convegno, ENG
Sannino, Giovanna; Falco, Ivanoe De; Pietro, Giuseppe De
Blood pressure is a basic physiological parameter in the cardiovascular circulatory system. PhotoPlethysmoGraphy (PPG) signal represents a convenient, wearable, and low-cost technology that can be applied to various aspects of cardiovascular monitoring, including the detection of blood pressure, i.e., the hypertension level. The goal of this paper is to explore the behavior of a set of machine learning methods with respect to the hypertension risk stratification. With reference to this issue, the discrimination ability of the investigated algorithms has been considered at three different granularity levels, i.e. by suitably joining some of the classes making up the data set, so as to partition this latter into three different ways. To fulfill our goal, the Cuff-Less Blood Pressure Estimation Data Set is considered here. This data set is composed by many signals, including PPG data acquired from a group of subjects, and their Blood Pressure values, that are used to represent their hypertension levels. We have used a large group of machine learning tools, relying on differing working methods, and their numerical comparison has been carried out in terms of risk stratification results.
2020, Editoriale in rivista, ENG
de Albuquerque V.H.C.; Gupta D.; De Falco I.; Sannino G.; Bouguila N.
Applied soft computing (Print) 95, pp. 1–52020, Articolo in rivista, ENG
Giovanna Sannino, Ivanoe De Falco and Giuseppe De Pietro
One of the most important physiological parameters of the cardiovascular circulatory system is Blood Pressure. Several diseases are related to long-term abnormal blood pressure, i.e., hypertension; therefore, the early detection and assessment of this condition are crucial. The identification of hypertension, and, even more the evaluation of its risk stratification, by using wearable monitoring devices are now more realistic thanks to the advancements in Internet of Things, the improvements of digital sensors that are becoming more and more miniaturized, and the development of new signal processing and machine learning algorithms. In this scenario, a suitable biomedical signal is represented by the PhotoPlethysmoGraphy (PPG) signal. It can be acquired by using a simple, cheap, and wearable device, and can be used to evaluate several aspects of the cardiovascular system, e.g., the detection of abnormal heart rate, respiration rate, blood pressure, oxygen saturation, and so on. In this paper, we take into account the Cuff-Less Blood Pressure Estimation Data Set that contains, among others, PPG signals coming from a set of subjects, as well as the Blood Pressure values of the latter that is the hypertension level. Our aim is to investigate whether or not machine learning methods applied to these PPG signals can provide better results for the non-invasive classification and evaluation of subjects' hypertension levels. To this aim, we have availed ourselves of a wide set of machine learning algorithms, based on different learning mechanisms, and have compared their results in terms of the effectiveness of the classification obtained.
DOI: 10.3390/jsan9030034
2020, Contributo in volume, ENG
Giovanna Sannino, Giuseppe De Pietro, Laura Verde
This chapter explores the most relevant aspects in relation to the outcomes and performance of the different components of a healthcare system with a particular focus on mobile healthcare applications. In detail, we discuss the six quality principles to be satisfied by a generic healthcare system and the main international and European projects, which have supported the dissemination of these systems. This diffusion has been encouraged by the application of wireless and mobile technologies, through the so-called m-Health systems. One of the main fields of application of an m-Health system is telemedicine, for which reason we will address an important challenge encountered during the realization of an m-Health application: the analysis of the functionalities that an m-Health app has to provide. To achieve this latter aim, we will present an overview of a generic m-Health application with its main functionalities and components. Among these, the use of a standardized method for the treatment of a massive amount of patient data is necessary in order to integrate all the collected information resulting from the development of a great number of new m-Health devices and applications. Electronic Health Records (EHR), and international standards, like Health Level 7 (HL7) and Fast Healthcare Interoperability Resources (FHIR), aims at addressing this important issue, in addition to guaranteeing the privacy and security of these health data. Moreover, the insights that can be discerned from an examination of this vast repository of data can open up unparalleled opportunities for public and private sector organizations. Indeed, the development of new tools for the analysis of data, which on occasions may be unstructured, noisy, and unreliable, is now considered a vital requirement for all specialists who are involved in the handling and using of information. These new tools may be based on rule, machine or deep learning, or include question answering, with cognitive computing certainly having a key role to play in the development of future m-Health applications.
2020, Articolo in rivista, ENG
De Falco, I., De Pietro, G., Sannino, G.
Automatic detection of falls is extremely important, especially in the remote monitoring of elderly people, and will become more and more critical in the future, given the constant increase in the number of older adults. Within this framework, this paper deals with the task of evaluating several artificial intelligence techniques to automatically distinguish between different activities of daily living (ADLs) and different types of falls. To do this, UniMiB SHAR, a publicly available data set containing instances of nine different ADLs and of eight kinds of falls, is considered. We take into account five different classes of classification algorithms, namely tree-based, discriminant-based, support vector machines, K-nearest neighbors, and ensemble mechanisms, and we consider several representatives for each of these classes. These are all the classes contained in the Classification Learner app contained in MATLAB, which serves as the computational basis for our experiments. As a result, we apply 22 different classification algorithms coming from artificial intelligence under a fivefold cross-validation learning strategy, with the aim to individuate which the most suitable is for this data set. The numerical results show that the algorithm with the highest classification accuracy is the ensemble based on subspace as the ensemble method and on KNN as learner type. This shows an accuracy equal to 86.0%. Its results are better than those in the other papers in the literature that face this specific 17-class problem.
2019, Articolo in rivista, ENG
Muhammad, Khan; Hussain, Tanveer; Tanveer, Muhammad; Sannino, Giovanna; de Albuquerque, Victor Hugo C.
Video summarization (VS) has attracted intense attention recently due to its enormous applications in various computer vision domains, such as video retrieval, indexing, and browsing. Traditional VS researches mostly target at the effectiveness of the VS algorithms by introducing the high quality of features and clusters for selecting representative visual elements. Due to the increased density of vision sensors network, there is a tradeoff between the processing time of the VS methods with reasonable and representative quality of the generated summaries. It is a challenging task to generate a video summary of significant importance while fulfilling the needs of Internet of Things (IoT) surveillance networks with constrained resources. This article addresses this problem by proposing a new computationally effective solution through designing a deep CNN framework with hierarchical weighted fusion for the summarization of surveillance videos captured in IoT settings. The first stage of our framework designs discriminative rich features extracted from deep CNNs for shot segmentation. Then, we employ image memorability predicted from a fine-tuned CNN model in the framework, along with aesthetic and entropy features to maintain the interestingness and diversity of the summary. Third, a hierarchical weighted fusion mechanism is proposed to produce an aggregated score for the effective computation of the extracted features. Finally, an attention curve is constituted using the aggregated score for deciding outstanding keyframes for the final video summary. Experiments are conducted using benchmark data sets for validating the importance and effectiveness of our framework, which outperforms the other state-of-the-art schemes.