Skip to main content

Detection of Alzheimer and mild cognitive impairment patients by Poincare and Entropy methods based on electroencephalography signals

Abstract

Alzheimer’s disease (AD) is characterized by deficits in cognition, behavior, and intellectual functioning, and Mild Cognitive Impairment (MCI) refers to individuals whose cognitive impairment deviates from what is expected for their age but does not significantly interfere with daily activities. Because there is no treatment for AD, early prediction of AD can be helpful to reducing the progression of this disease. This study examines the Electroencephalography (EEG) signal of 3 distinct groups, including AD, MCI, and healthy individuals. Recognizing the non-stationary nature of EEG signals, two nonlinear approaches, Poincare and Entropy, are employed for meaningful feature extraction. Data should be segmented into epochs to extract features from EEG signals, and feature extraction approaches should be implemented for each one. The obtained features are given to machine learning algorithms to classify the subjects. Extensive experiments were conducted to analyze the features comprehensively. The results demonstrate that our proposed method surpasses previous studies in terms of accuracy, sensitivity, and specificity, indicating its effectiveness in classifying individuals with AD, MCI, and those without cognitive impairment.

Introduction

Alzheimer’s disease (AD) is probably the most widespread neurodegenerative disorder affecting millions of individuals worldwide, and approximately 50 million patients are predicted to suffer from this disease [1]. AD can also lead to dementia. Although earlier diagnosis and detection can help to prevent the progression of dementia and thus improve the quality of life of patients, there is no definite cure for AD. In terms of dementia, mild cognitive impairment (MCI) refers to patients who are not severe enough to meet AD diagnostic criteria but have cognitive deficits that are greater than expected for their age. There is the possibility that the MCI patients may progress to AD over time. Therefore, the diagnosis of MCI patients in earlier stages could be crucial in case of preventing the progression of disease.

There are several ways to detect people who suffer from neurodegenerative diseases. Clinically, the main method for identifying AD and MCI patients is a subjective neuropsychological test. In addition, in the case of AD and MCI disorders, analyzing changes in the structure of the brain has been highlighted by researchers. Several studies have been proposed to detect and analyze AD patients based on MR images [2,3,4].

Although imaging systems could yield better results in terms of localization and detection of any abnormalities in the brain, owing to the high cost and time consumption of these procedures, electroencephalography (EEG) is generally used to detect changes and abnormalities in brain activity, since it is noninvasive, fast and inexpensive with high accuracy results. EEG has been proven to be an effective tool for characterizing several disorders, such as sleep disorders [5], epilepsy detection [6, 7], stress disorder, and detecting alcoholism [8]. Therefore, with the aim of developing biomarkers for early detection, diagnosis, and monitoring of disease progression in AD and MCI patients, EEG signal processing methods have been utilized for several decades.

In the literature, numerous studies have been proposed to analyze the EEG signals of AD and MCI patients. To evaluate the EEG signals, the pre-processing step can be implemented to eliminate noise and unwanted data. Since EEG signals have low-frequency bands, an appropriate low-pass filter and wavelets-based filters can be designed [9, 10]. Although combinations of methods have been proposed to examine EEG signals, approaches can generally be categorized into four groups:

  • Time domain: Evaluations of EEG signals using time-domain features such as waveform length, zero-crossing and mean absolute value, skewness, and kurtosis were presented in the literature [11,12,13]

  • Frequency domain: To analyze characteristics of EEG signals from individuals with MCI, AD, and healthy individuals in the frequency domain, Fourier transform (FT) and short-time FT are used to extract meaningful features [14,15,16,17].

  • Time–frequency domain: Since brain activity occurs in low-frequency bands, analyzing the changes in frequency bands is essential in EEG signal processing, in which wavelets are the most used approach for accessing these frequency bands [18,19,20].

  • Non-linear methods: For non-linear biomedical signal processing, several methods, including Entropy, Poincare, and Lyapunov exponent analysis, are used to evaluate the complexity of signals and the stability of any steady-state behavior [21,22,23,24].

Since brain activity is not linear, non-linear methods could be used to evaluate EEG signals. It has been presented by [25,26,27] that to characterize EEG signals in different circumstances, non-linear methods are beneficial, especially if the amount of data is relatively high. In the literature, few studies focus on classifying three groups: AD, MCI, and healthy individuals. The main objective of this study is to apply machine learning techniques to non-linear features extracted using widely adopted Poincaré and Entropy methods for the classification of these three groups. The proposed method not only addresses the research gap in classifying these three groups but also achieves the highest classification accuracy reported in studies using EEG signals. In addition, the classification between AD-MCI, AD-healthy, and healthy-MCI has been examined.

Related works

Among the broad spectrum of research dedicated to diagnosing Alzheimer’s Disease (AD) and mild cognitive impairment (MCI) using electroencephalography (EEG) signals, this study has conducted a literature review with a specific emphasis on nonlinear methods of analysis.

Several studies have employed entropy-based methods to analyze EEG signals for diagnosing AD and MCI. In [28], the authors used Permutation Entropy (PE) to analyze EEG signals in 2-s epochs, revealing that the complexity of MCI and AD patients is lower than that of healthy individuals, with better classification in the eye-open state. Similarly, [29] applied Dispersion Entropy on sub-frequency bands, followed by an ANOVA-based dimension reduction technique. Spectral Entropy (SpecEn) was introduced in [30] for analyzing EEG signals of AD, MCI, and healthy individuals, contributing to a better understanding of spectral characteristics in these groups.

Wavelet and Approximate Entropy approaches have also been explored in the literature. In [31], the authors investigated the relationship between Wavelet Entropy and power distribution across EEG channels, with the Haar wavelet showing the best performance in distinguishing AD patients from healthy individuals. Similarly, [32] applied Approximate Entropy and Average Mutual Information to classify AD and healthy individuals, achieving an AUC of 0.91. In addition, Chai et al. [33] analyzed sub-frequency bands and multi-scale entropy across 25-s EEG epochs, achieving an AUC of 0.89 in AD detection.

Several nonlinear-based analyses have also been applied to EEG signals for AD diagnosis. Staudinger and Polikar [34] conducted a non-linear analysis of EEG signals from AD and healthy patients. Houmani et al. [35] introduced an epoch-based approach using Shannon Entropy combined with the Hidden Markov Model (HMM) for early diagnosis of AD. Furthermore, Tsallis Entropy was employed in [36] to detect Alzheimer’s disease, and Siuly et al. [37] proposed a combination of Permutation Entropy and autoregression models to identify MCI patients. A comprehensive analysis of sub-frequency bands of all groups was performed by Tzimourta et al. [27], in which statistical time analysis and non-linear methods were employed for each band. In addition, the authors indicated each band’s effectiveness in different brain regions.

Research into the use of entropy-based and optimization algorithms has expanded the potential for EEG-based classification. Simons et al. [38] applied quadratic sample entropy across EEG channels, achieving 78% accuracy in detecting AD. Nonlinear methods, including Singular Value Decomposition, Detrended Fluctuation Analysis, and Higuchi Dimension, were employed to analyze Alzheimer’s disease based on EEG signals, as presented in [39]. The obtained results highlight the effectiveness of the SVD method in achieving high-quality features with almost 93% accuracy in detecting AD and healthy adults. In [40], Multi-scale Fuzzy Entropy was applied to sub-frequency bands to identify AD and Control groups. Although relatively low accuracy was achieved in comparison with other studies, implementing just one algorithm is the strength of the approach. Optimal selection of EEG channels by Intelligent Optimization algorithm with focusing on entropy and the Second Order Difference Plot (SODP) features proposed by [41]. Researchers identified FP1, F3, F4, and P4 channels as most effective for distinguishing between Alzheimer’s disease and healthy controls, achieving high classification accuracy and sensitivity. Combination of several entropy-based feature extraction methods have been considered in [42]. Analysis of different epoch lengths was presented, which showed that the entropy values of the MCI group for lower epoch periods(1 ≤ s ≤ 4) were lower than the control group, and for longer epoch lengths (10 ≤ s ≤ 15), the entropy values of the MCI group were higher than the control group. In addition, [43] used a combination of multi-domain features and feature selection methods to classify motor imagery EEG signals, proposing a novel approach to enhance classification accuracy.

The primary contribution of this study lies in developing a new computer-based decision support system for the discrimination of AD, MCI, and healthy subjects. Specifically, the key contributions of this study include the application of nonlinear methods in EEG signal analysis, the evaluation of features obtained from each method, and the determination of each group using machine learning algorithms.

Data set

The EEG signals used in this study were gathered at the Neurology Department of Baskent University Hospital, following the guidelines set by the Baskent University Institutional Review Board and Ethics Committee [14]. A total of 35 participants, aged between 65 and 90, participated in the study. The data sizes of the groups were assessed for conformity using “One ROC Curve Power Analysis”. The hypothesis suggests that with 80% power and a 5% error rate, the true detection rate could reach approximately 90%. Demographic characteristics of the participants is given in Table 1.

Table 1 Overview of participants

Every participant underwent a thorough examination to rule out the existence of additional neurological disorders, such as epilepsy, multiple sclerosis, and Parkinson’s disease, as well as depression or other psychiatric disorders. In addition, degenerative dementia syndromes, including frontotemporal dementia, Lewy body dementia, and vascular dementia, were checked, as well as the presence of systolic or uncontrolled hypertension, substance or alcohol abuse, a history of traumatic brain injury, and structural lesions in the brain were also considered and excluded.

19 EEG electrodes were replaced based on the 10–20 international system (Fig. 1), and during resting eyes-closed condition, 7 min of EEG data were recorded using a Nihon Kohden-Neurofax EEG 1200. The sampling frequency and bit resolution are 200 Hz and 16 bits, respectively.

Fig. 1
figure 1

10–20 system electrode montage

Although a variety of epoch lengths could be considered (as given in the last section, Table 8), an epoch length of 30 s was considered in this study. Accordingly, 14 epochs were examined for each channel, resulting in 490 epochs.

Methodology

This study aims to employ nonlinear methods to extract features from EEG signals. As given in Fig. 2, the proposed method includes 2 steps; in the first step, several methods are employed to extract meaningful features from the EEG signal. The obtained feature set is given to machine learning algorithms in the second step for examination. The performance of the proposed approach is determined by evaluating the test set. Pseudocode of the proposed method is given in Table 2. Two different methods were implemented to extract features. A brief description of them is given below.

Fig. 2
figure 2

Proposed method’s architecture

Table 2 Pseudocode of proposed method

Entropy

Although Entropy is used in various fields, in nonlinear dynamical systems, entropy can be implemented to quantify the complexity of a system. Since EEG signals are non-stationary and nonlinear, entropy is widely used to evaluate the complexity of signals [29,30,31,32,33,34,35,36,37,38]. In this regard, several entropy-based methods were utilized in this study to extract features from each epoch, as discussed below.

Permutation entropy (PerEn)

To examine the complexity or randomness of the time series, the PerEn determines a certain permutation pattern in the signal [44]. A given time series is first transformed into a series of ordinal patterns in which each pattern specifies the relationship between the current and past sequences. PerEn is defined as follows:

$$\text{PerEn}= - {\sum }_{j=1}^{k!}{p}_{j}{\text{log}}_{2}({p}_{j})$$
(1)

where k is the permutation order, and \({p}_{j}\) is the relative frequency of the permutation pattern. (k is chosen 3 in this study).

Approximate entropy (AppEn)

Given a time series x = {x1, x2, x3, … xN} in m dimensional phase-space, AppEN states the likelihood of similar patterns of observations [45]. Thus, for high irregularity time series, AppEN would be high.

The similarity between pairs is calculated by \(\text{d }({x}_{m}\left(\text{i}\right), {x}_{m}\left(\text{j}\right))\) which is the maximum absolute difference between their respective scalar components, i.e., the maximum norm:

$$d\left[X\left(i\right), X\left(j\right)\right]=\underset{k=\mathit{ }\text{1,2},\dots ,m}{\text{max}}\left|x\left(i+k-1\right)-x(j+k-1)\right|$$
(2)

For each i = 1 i N m C 1, the correlation integral Cmi can be formed as

$${C}_{r}^{m}\left(r\right)=\frac{1}{N-m+1} \sum_{j=1}\uptheta (\text{r}-\text{d }({x}_{m}\left(\text{i}\right), {x}_{m}(\text{j}))$$
(3)

where θ is Heavyside function (θ(x) = 1 for x > 0, θ(x) = 0, otherwise]). Then, φ is defined as

$$\varphi^{m} \left( r \right) = \frac{1}{N - m + 1} \mathop \sum \limits_{i = 1}^{N - m + 1} \ln C_{r}^{m} \left( i \right)$$
(4)

This procedure continues for the m + 1 dimension, and AppEn is defined as follows:

$$\text{AppEn}\left(m,r,\uptau , N\right)= {\uptheta }^{m}\left(r\right)- {\uptheta }^{m+1}(r)$$
(5)

where m is the number of phase space dimensions, r is the tolerance, N is the time series length, and τ is lag.

Sample entropy (SamEn)

Sample entropy is a modified version of AppEn that outperforms in the independence of length of the time series as well as trouble-free implementation and is defined as [46]

$$\text{SamEn}= - \text{log}(\frac{A}{B})$$
(6)

where A and B are the total number of vector pairs having d(x m+1 (i), xm+1(j)) < r of length (m + 1) and the total number of vector pairs having d(xm(i), xm(j)) < r of length m, respectively.

Spectral entropy (SpecEn)

The measure of unpredictability and randomness of a time series by SpecEn is associated with the power spectrum of the components of a signal [47]. SpecEn is high if the power of signals has a wide range of frequencies:

$${H}_{2}(X)= -\text{log}\sum_{i=1}^{n}{p}_{i}^{2}$$
(7)

where pf is the power spectral density Fourier transform at frequency f.

Bubble entropy (bEn)

A modification of permutation entropy was given by [48] as Bubble Entropy. This new perspective is free of parameters m and r. For each vector in m-dimensional phase space, the Renyi entropy is calculated by

$${H}_{2}(X)= -\text{log}\sum_{i=1}^{n}{p}_{i}^{2}$$
(8)

and bEn is achieved by the Renyi entropy of successive dimensions divided by normalization factor log (m + 1/m-1):

$${\text{bEn}} = \left( {H_{{{\text{swaps}}}}^{m + 1} - H_{{{\text{swaps}}}}^{m} } \right)/\log \left( {m + 1/m - 1} \right)$$
(9)

Entropy of entropy (EoE)

EoE computes Shannon entropy twice for a signal utilizing a multiscale approach [49]. EoE consists of two steps; in each step, the Shannon entropy is implemented into a time series. First time series x = {x1, x2, …., xN,} is divided into several consecutive nonoverlapping windows w = {x(j-1)τ+1,…, x(j-1)τ+τ}, then Shannon Entropy is implemented into each window. The probability of certain patterns is calculated by

$${p}_{jk}= \frac{\text{total number of} {x}_{j}\text{over} {w}_{j}^{\tau }\text{in} \text{state} k}{\tau }$$
(10)

where k is the state index from 1, and Shannon entropy is consequently implemented as follows:

$${y}_{j}^{\tau }= -\text{log}\sum_{k=1}^{{s}_{1}}{p}_{jk}\text{ln}{p}_{jk}$$
(11)
$${y}_{j}^{\tau }= -\text{log}\sum_{k=1}^{{s}_{1}}{p}_{jk}\text{ln}{p}_{jk}$$

In addition, consequently, EoE is computed under different time scales τ:

$$EoE= -\sum_{l=1}^{{s}_{2}}{p}_{l} ln{p}_{l}$$
(12)

Wavelet entropy (WE)

For non-stationary signals, wavelet entropy combines wavelet decomposition with the order within wavelet coefficients [50]. Wavelet Entropy calculates the entropy of the normalized wavelet coefficients by considering them as an empirical probability distribution:

$$WE= \sum_{i}\sum_{j}{\left.\left|{wt}_{ij}\right.\right|}^{2}$$
(13)

where the probability mess function is

$$P\left({wt}_{ij}\right)= {\left.\left|{wt}_{ij}\right.\right|}^{2}IE$$
(14)

Fuzzy entropy (FuzzEn)

Fuzzy entropy (FuzzEn) provides robust estimates of signal uncertainties, even in the presence of noise, and is highly responsive to changes in information content [51]. By analyzing short signals, FuzzEn can achieve reliable and stable estimations. The similarity degree between two vectors xm(i) and xm(j) can be defined as

$${D}_{ij}^{m}= \mu \left({d}_{ij}^{m}, n, r\right)= {e}^{\frac{{{- (d}_{ij}^{m})}^{n}}{r}}$$
(15)

where dijm is the distance between two vectors xm(i) and xm(j), and μ is the fuzzy membership function. θ is defined as

$${\uptheta }^{m}\left(n,r\right)=\frac{1}{N-m} \sum_{1}^{N-m}(\frac{1}{N-m-1} \sum_{j=1, j\ne 1}^{N-m}{D}_{ij}^{m})$$
(16)

Therefore, FuzzEn for time series is obtained by

$$\text{FuzzEn}\left(m,n,r,N\right)=\text{ln}{\uptheta }^{m}\left(n,r\right)-\text{ln}{\uptheta }^{m+1}(n,r)$$
(17)

where r is similarity tolerance which was chosen 0.25 times the standard deviation of the signal. m, n and N are embedding dimension, number of datapoints and length of the time series, respectively.

Disperse entropy (DispEn)

To quantify the regularity of the time series, DispEn was introduced by [52], which uses the normal cumulative distribution and maps the original signal to the dispersion signal. Therefore, higher complexity results in higher DispEn value.

First, a mapping function is implanted to normalize the time series:

$${y}_{f}= \frac{1}{\sigma \sqrt{2\pi }} {\int }_{-\infty }^{x}{e}^{\frac{{-(s-\mu )}^{n}}{{2\sigma }^{2}}} ds$$
(18)

where Ϭ is the variance, and μ is the expectation of the normal data distribution. Normalized data are defined as follows:

$${y}_{f}^{m}=\left[{y}_{f}+ {y}_{f+\tau }, \dots , {y}_{f+(m-1)\tau }\right]$$
(19)

where j = 1, 2, …, n − (m − 1)τ, m is phase space dimension and τ is time lag. Then, ym is grouped into c classes:

$${z}_{i}^{c}=\text{round}(c. {y}_{f}+0.5)$$
(20)
$${z}_{f}^{m, c}=[{z}_{f}^{c}, {z}_{f+\tau }^{c},\dots ,{z}_{f+(m-1)\tau }^{c}]$$
(21)

and the relative frequency of each dispersion pattern is achieved by

$$P\left(\pi \right)= \frac{\text{Number} \{\pi \}}{N-(m-1)\tau } (1\le i\le {c}^{m})$$
(22)

where Pi) represents the ratio of the number of ith dispersion patterns to the number of embedding vectors, finally, based on Shannon entropy, DispEn is defined as

$$\text{DispEn} \left(m, c,\uptau ,N\right)= - \sum_{\pi =1}^{{c}^{m}}P\left(\pi \right).\text{ln}(P\left(\pi \right))$$
(23)

where m is the number of dimensions of the phase space, and τ is lag. N and C are number of classes and length of time series, respectively.

Attention entropy (AttEn)

To overcome the disadvantage of earlier entropy methods, which based on counting the frequency of all observations, AttEn was presented by analyzing the frequency distribution of the intervals between the critical observations in a time series [53]. In this method, first local minimum and maximum are computed and then create sets of intervals between these minimums and maximums. Then, the probability of each element appearing in the set based on the number of appearances is calculated by applying Shannon entropy:

$$\text{AttEn}(f)= -\sum_{l}f \text{ln}({f}_{l})$$
(24)

Poincare plot

To assess the qualitative characteristics of biological signal patterns, the phase space technique can be used. Several methods have been presented thus far, and one of the most commonly used methods is the Poincare plot. A Poincare plot is a nonlinear method in analyzing time series and represents the geometrical pattern of a chaotic system’s dynamics. In the biomedical signal evaluation and feature extraction, a Poincare plot could help understand the correlation of data points [54,55,56]; for example, in ECG signal analysis, a Poincare plot represents the dynamics of Heart Rate Variability (HRV) [57]. Visualizing one-dimensional biomedical signals and utilizing image processing techniques are crucial in medical diagnosis and treatment planning. Geographical feature analysis, which was implemented in several studies, proposed to extract meaningful features from Poincare and phase-space followed by artificial intelligence and machine learning tools to analyze the intrinsic behavior of diseases, such as depression, recognition of seizure and alcoholism disorders [8, 58, 59].

The time lag should be specified beforehand to examine the Poincare plot of the time series. Therefore, serial dependence of time series could be examined via predictions from previous observations. The lag is generally set to 1 in a typical Poincare plot, so the signal is plotted vs. its lag delay. As depicted in Fig. 3, the x-axis is the time series (here it is one epoch), and the y-axis is the lagged time series. Five different features are extracted from the plot as follows: (1) SD1: the standard deviation of the points calculated along the direction perpendicular to the line 1; (2) SD2: the standard deviation of the points along the line 2; (3) the rate of SD1/SD2; (4) the rate of SD2/SD1; and (5) the area of the ellipse.

Fig. 3
figure 3

Poincare plot’s characteristics of a simple EEG signal

Classification

Machine learning algorithms are generally used to make predictions and analyze the data set [59]. This study used several methods to properly examine data set and obtained features. In the following, a short description of methods is given.

K-nearest neighbor

K-nearest neighbor (KNN) classifier makes prediction based on the distance of test data set with training data. Based on predefined values for k, it measures the new sample instance with training instances around it, and using majority voting, it assigns a label to the test instance.

Support vector machine

SVM (Support Vector Machine) operates by identifying the optimal hyperplane. Objective of SVM classifier is to locate the maximum margin hyperplane which exhibits the most significant distance between the support vectors, which are the closest data points in the space.

Bagged tree

The bagging algorithm gathers the outputs of the base classifiers, which are trained on instances and combines them by voting for the labels to make predictions.

  

Actual Labels

AD

Control

MCI

Predicted Labels

AD

A

B

C

Control

D

E

F

MCI

G

H

I

Measures

FN

B + C

D + F

G + H

TN

E + F + H + I

A + C + G + I

A + B + D + E

FP

D + G

B + H

C + F

Performance metrics

To evaluate the original data set, the physician’s point of view about the diagnosis was considered standard. To examine the performance of classifiers, several measures were considered. Since there are three classes, the confusion matrix can be defined. The sensitivity, specificity, and accuracy are mainly used metrics that can provide information about the outcomes of machine learning algorithms. These measures are defined as follows:

$$\text{Accuracy}= \frac{\text{True Positive}+\text{True Negative} }{\text{All possibilities}}$$
(25)
$$\text{Sensitivity}= \frac{\text{True Positive} }{\text{True Positive}+\text{False Negative}}$$
(26)
$$\text{Specificity} = \frac{\text{True Negative} }{\text{True Negative}+\text{False Positive}}$$
(27)

Experiments and results

Parameters

To better examine the data set and avoid overfitting, in this study, K-fold cross validation was considered. The number of folds was set to 5. In addition, the test-train ratio was selected as 0.3, and none of the subjects in the train set were not used for testing. after training, the model was tested using an unused test set. For the SVM classifier, the kernel function was chosen as Cubic and Quadric SVM with box constraint level set to 1, and kernel scale mode was selected as “auto”. Number of neighbors was chosen as 1 and 3 for KNN. The parameters of Bagged Trees were adjusted as follows: the maximum number of splits was set to 489, and the number of learners was set to 30. Furthermore, for entropy methods, phase space dimension and tolerance rate were chosen as 2 and 0.25 times the standard deviation of signal, respectively. C in DispEn is set to 6. In addition, the time lag for both the Poincare and Entropy approaches set to 1.

Results

Since two different nonlinear tools (Entropy and Poincare) were used in this study, three different experiments were considered to have a better examination of the features obtained. Achieved features by Entropy and Poincare methods were analyzed separately. For each of these experiments, the related method was employed, and features were given to machine learning classifiers. In the next experiment, all features were examined by implementing both Poincare and Entropy approaches. As there are 14 epochs and 19 channels for each patient, the feature matrix for the Poincare (5 features) and Entropy (10 features) methods are 490 × 95 and 490 × 190, respectively. In addition, for the last experiment, total of 15 feature extraction methods were implemented, which generated a feature matrix with a size of 490 × 285. Moreover, since the results of machine learning algorithms depend on the initial conditions and parameters, each algorithm was run 5 times, and the mean of all experiments was considered as the final result.

An example of the Poincare plot for one epoch of EEG data is given in Fig. 4. In this example, electrodes F3–C3 and P3–O1 are depicted for one sample of each group. The x-axis is the time series, while the y-axis is one lagged of that series. As it is obviously seen, there is a geometrical difference between groups. Therefore, in the case of statistical properties, distinguishing between groups would also be straightforward.

Fig. 4
figure 4

From left to right: Control, AD and MCI for A F3–C3 and B P3–O1 channels

Table 3 shows the evaluation of different feature sets and classifiers. Results of both training and test sets are given. To analyze the effectiveness of features, three different feature sets were constructed. As indicated by the Poincaré plot, the statistical features derived from this plot play a crucial role, achieving an accuracy of 91.9% in classifying the three groups in the test set. For entropy features, SVM and KNN classifiers performed better than the Poincare method. However, if all features are taken into account, a better performance was obtained, which is 97.8% in accuracy, using the SVM classifier with a cubic kernel function.

Table 3 Classification accuracy of the Control, MCI, and AD groups

In addition, since most studies aimed to distinguish between two groups, classification between AD vs. Control, AD vs. MCI, and Control vs. MCI groups was also performed. It was assumed that distinguishing between AD and Healthy individuals would be straightforward, owing to distinct differences between characteristics of EEG signals. As given in Table 4, machine learning algorithms achieved excellent results in distinguishing between groups in which, by considering all features, 98.7% accuracy was obtained using the KNN classifier.

Table 4 Classification accuracy of the AD and Control groups

On the other hand, classification between Control and MCI was more challenging than in previous experiment. Results of the classification of experiments are given in Tables 5 and 6, wherein the best situation, the accuracy is 97.2% and 98.2% for Control vs. MCI and AD vs. MCI, respectively.

Table 5 Classification accuracy of Control and MCI groups
Table 6 Classification accuracy of the AD and MCI groups

Since the SVM cubic classifier achieved successful results, the sensitivity and specificity metrics for test sets for the SVM cubic classifier are given in Table 7. In the classification of 3 groups, in case all features were considered, sensitivity and specificity were 98.6% and 98.2%, respectively, which are more significant than using each method separately.

Table 7 Sensitivity and Specificity metrics for test set in SVM classifier with Cubic kernel function

Discussion

The proposed computer-based decision support system based on Entropy and Poincare features can discriminate all groups with high accuracy. Based on the results obtained in the previous section, entropy features proved particularly powerful, although the Poincaré method also yielded acceptable results. However, by considering all features, more reliable systems with better performance were designed. Our strategy to detect all groups was successful, and an accuracy of 97.8% was achieved by the SVM classifier. This surpasses the 96.5% accuracy achieved in our previous study, which required more complex strategies, such as preprocessing and discrete wavelet transform (DWT) [14]. In addition, 98.6% and 98.2% accuracy, respectively, were obtained for the test set. According to the classifier evaluation, although the KNN and SVM results are similar, the SVM classifier with the cubic kernel function performed slightly better than the other classifiers. Since many related studies have been published in this field, we aimed to compare the obtained results with those of other studies in which nonlinear methods were used.

As given in Table 8 and discussed in Sect. “Related works”, various nonlinear methods with different parameters have been considered in this study. While many studies classify between two groups (e.g., AD and MCI or AD and Control), our method efficiently handles multiclass classification. For a nonlinear dynamical system, the duration of an event is a significant issue and should not be chosen too long or short. Most of the studies consider shorter epoch lengths, which contributes to more computations. For the classification of all groups in literature, based on permutation entropy, Şeker et al. [28] suggested 100% in F1 score, but for better comparison, additional evaluation metrics are needed, such as sensitivity or accuracy. In addition, a limited length of epochs was considered, which increased calculation costs. Among the remaining methods, our proposed method achieved the highest accuracy rate.

Table 8 Comparison of the results of the studies in the literature and the proposed methodology

In the classification between AD and MCI patients, our proposed method achieved 98.2% accuracy when using the SVM classifier in case all features were considered. In addition, features obtained by Poincare and Entropy achieved 95.5 and 96.5, respectively, highlighting the importance of features. Amezquita-Sanchez et al. [29] achieved 91% using the dispersion entropy index, while the length of the records is shorter (4 min).

In the case of AD and Control classification, the proposed method acquires 98.7% accuracy while considering all features with the SVM classifier. The proposed model outperforms others, while [29] and [35] achieved approximately 97% and 90% accuracy, respectively. The researchers in [39] achieved an accuracy of 94.72% in distinguishing between AD and Control groups using the SVD feature and kNN classifier. In addition, the application of Auto Mutual Information and Approximate Entropy for classifying AD and Control groups, as proposed by [32], resulted in an accuracy of 91.74%.

Finally, in the case of MCI and Control groups, the proposed method obtained 97.2% accuracy by Entropy features with SMV classifier. Using an auto-regressive model and permutation entropy (PE), Siuly et al. [37] achieved slightly better results in this case. Despite the simplicity method presented by [40] which is using a single entropy measure, the method’s accuracy was 83% which is notably lower than this study, suggesting that this approach may lack robustness and precision in differentiating between healthy and AD individuals. Authors in [41] achieved high diagnostic performance with an accuracy of 93.53%, sensitivity of 98.74% and specificity of 98.25% indicating its robustness in pattern recognition and potential to reduce misdiagnosis. Nevertheless, the computational complexity and reliance on high-quality EEG data may constrain its real-time applicability and generalizability across diverse populations.

Finally, it is essential to note that most EEG signal processing approaches involve filtering to isolate desired frequencies and remove noise. However, this study applied the proposed methodology to raw data without any preprocessing steps.

Conclusions

Alzheimer disease is the most common neurodegenerative disorder in the world. With no specific cure available, early detection is vital for mitigating disease progression through timely intervention. Since the diagnosis and early detection of these diseases are challenging, the proposed computer-assisted method can aid in analyzing the progression of these disorders. This paper evaluated the effectiveness of nonlinear processes in feature extraction from electroencephalography (EEG) records of individuals with AD, MCI, and healthy controls. Two nonlinear methods (Entropy and Poincare) were implemented on the EEG records of AD, MCI, and healthy individuals. The features extracted were then analyzed using various machine learning algorithms to distinguish between the three groups. For the 3-class classification, an accuracy of 97.6% was achieved, while the sensitivity and specificity rates were 98.6% and 98.2%, respectively.

Based on the results obtained, all three groups achieved an acceptable classification rate. When examining the Poincare plot, it was observed that the pattern of each group was different from that of other groups, and based on the statistical analysis of extracted features from the Poincare method, successful results were achieved. On the other hand, the performance of the entropy-based method was better than the Poincare method. The proposed methodology offers a promising alternative to the more expensive and time-consuming MR imaging techniques currently in use. The high rates of accuracy, sensitivity, and specificity achieved indicate the potential of these combined nonlinear methods in facilitating the early detection and diagnosis of AD and MCI.

Although diagnosing Alzheimer’s using MRI image processing approaches is more effective and has a higher accuracy rate, the proposed method makes diagnosis more effective and cheaper with an acceptable high accuracy rate. However, to generalize the reliability of the proposed approach, it is necessary to apply the method to a larger experimental data set. Hence, further investigation should incorporate the application of the method on a high number of objects to validate its robustness. In addition, upon reviewing previous studies, it becomes apparent that decision support systems have employed a limited array of Entropy methods. Future inquiries endeavor to enhance the efficacy of the proposed method by judiciously selecting appropriate entropy methods through the application of feature selection techniques.

Availability of data and materials

No data sets were generated or analysed during the current study.

References

  1. Prince M, Bryce R, Albanese E, Wimo A, Ribeiro W, Ferri CP. The global prevalence of dementia: a systematic review and metaanalysis. Alzheimers Dement. 2013;9(1):63–75.

    Article  Google Scholar 

  2. Feng J, Zhang SW, Chen L. Identification of Alzheimer’s disease based on wavelet transformation energy feature of the structural MRI image and NN classifier. Artif Intell Med. 2020;108: 101940.

    Article  Google Scholar 

  3. Nguyen M, He T, An L, Alexander DC, Feng J, Yeo BT, Initiative ADN. Predicting Alzheimer’s disease progression using deep recurrent neural networks. Neuroimage. 2020;222: 117203.

    Article  Google Scholar 

  4. Lee E, Choi JS, Kim M, Suk HI, Alzheimer’s Disease Neuroimaging Initiative. Toward an interpretable Alzheimer’s disease diagnostic model with regional abnormality representation via deep learning. Neuroimage. 2019;202: 116113.

    Article  Google Scholar 

  5. Liang SF, Kuo CE, Hu YH, Pan YH, Wang YH. Automatic stage scoring of single-channel sleep EEG by using multiscale entropy and autoregressive models. IEEE Trans Instrum Meas. 2012;61(6):1649–57.

    Article  Google Scholar 

  6. Sharmila A, Geethanjali P. Evaluation of time domain features on detection of epileptic seizure from EEG signals. Health Technol. 2020;10:711–22.

    Article  Google Scholar 

  7. Fürbass F, Hartmann M, Perko H, Skupch A, Dollfuß P, Gritsch G, et al. Combining time series and frequency domain analysis for a automatic seizure detection. In 2012 Annual International Conference of the IEEE Engineering in Medicine and Biology Society, 2012;1020–1023.

  8. Sadiq MT, Akbari H, Siuly S, Li Y, Wen P. Alcoholic EEG signals recognition based on phase space dynamic and geometrical features. Chaos, Solitons Fractals. 2022;158: 112036.

    Article  MathSciNet  Google Scholar 

  9. Sadiq MT, Yu X, Yuan Z, Aziz MZ. Motor imagery BCI classification based on novel two-dimensional modelling in empirical wavelet transform. Electron Lett. 2020;56(25):1367–9.

    Article  Google Scholar 

  10. Sadiq MT, Yu X, Yuan Z, Zeming F, Rehman AU, Ullah I, et al. Motor imagery EEG signals decoding by multivariate empirical wavelet transform-based framework for robust brain–computer interfaces. IEEE Access. 2019;7:171431–51.

    Article  Google Scholar 

  11. Ma D, Zheng J, Peng L. Performance evaluation of epileptic seizure prediction using time, frequency, and time–frequency domain measures. Processes. 2021;9(4):682.

    Article  Google Scholar 

  12. Cura OK, Yilmaz GC, Ture HS, Akan A. Classification of Alzheimers’ Dementia by Using Various Signal Decomposition Methods. In 2021 Medical Technologies Congress (TIPTEKNO), 2021; 1–4.

  13. Rad EM, Azarnoosh M, Ghoshuni M, Khalilzadeh MM. Diagnosis of mild Alzheimer’s disease by EEG and ERP signals using linear and nonlinear classifiers. Biomed Signal Process Control. 2021;70: 103049.

    Article  Google Scholar 

  14. Oltu B, Akşahin MF, Kibaroğlu S. A novel electroencephalography based approach for Alzheimer’s disease and mild cognitive impairment detection. Biomed Signal Process Control. 2021;63: 102223.

    Article  Google Scholar 

  15. Klepl D, He F, Wu M, De Marco M, Blackburn DJ, Sarrigiannis PG. Characterising Alzheimer’s disease with EEG-based energy landscape analysis. IEEE J Biomed Health Inform. 2021;26(3):992–1000.

    Article  Google Scholar 

  16. Miao Y, Jurica P, Struzik ZR, Hitomi T, Kinoshita A, Takahara Y, et al. Dynamic theta/beta ratio of clinical EEG in Alzheimer’s disease. J Neurosci Methods. 2021;359: 109219.

    Article  Google Scholar 

  17. Pirrone D, Weitschek E, Di Paolo P, De Salvo S, De Cola MC. Eeg signal processing and supervised machine learning to early diagnose alzheimer’s disease. Appl Sci. 2022;12(11):5413.

    Article  Google Scholar 

  18. Polikar R, Topalis A, Green D, Kounios J, Clark CM. Comparative multiresolution wavelet analysis of ERP spectral bands using an ensemble of classifiers approach for early diagnosis of Alzheimer’s disease. Comput Biol Med. 2007;37(4):542–58.

    Article  Google Scholar 

  19. Adeli H, Ghosh-Dastidar S, Dadmehr N. A spatio-temporal wavelet-chaos methodology for EEG-based diagnosis of Alzheimer’s disease. Neurosci Lett. 2008;444(2):190–4.

    Article  Google Scholar 

  20. Puri DV, Nalbalwar SL, Nandgaonkar AB, Gawande JP, Wagh A. Automatic detection of Alzheimer’s disease from EEG signals using low-complexity orthogonal wavelet filter banks. Biomed Signal Process Control. 2023;81: 104439.

    Article  Google Scholar 

  21. Chen S, Zhang X, Chen L, Yang Z. Automatic diagnosis of epileptic seizure in electroencephalography signals using nonlinear dynamics features. IEEE Access. 2019;7:61046–56.

    Article  Google Scholar 

  22. Hsu CH, Tsai MY, Huang GS, Lin TC, Chen KP, Ho ST, et al. Poincaré plot indexes of heart rate variability detect dynamic autonomic modulation during general anesthesia induction. Acta Anaesthesiol Taiwan. 2012;50(1):12–8.

    Article  Google Scholar 

  23. Chen X, Xu G, Du C, Zhang S, Zhang X, Teng Z. Poincaré plot nonextensive distribution entropy: a new method for electroencephalography (EEG) time series. Sensors. 2022;22(16):6283.

    Article  Google Scholar 

  24. Yakovleva TV, Kutepov IE, Karas AY, Yakovlev NM, Dobriyan VV, Papkova IV, et al. EEG analysis in structural focal epilepsy using the methods of nonlinear dynamics (Lyapunov exponents, Lempel–Ziv complexity, and multiscale entropy). The Scientific World J. 2020.

  25. Müller KR, Anderson CW, Birch GE. Linear and nonlinear methods for brain-computer interfaces. IEEE Trans Neural Syst Rehabil Eng. 2003;11(2):165–9.

    Article  Google Scholar 

  26. Lehnertz K, Mormann F, Kreuz T, Andrzejak RG, Rieke C, David P, Elger CE. Seizure prediction by nonlinear EEG analysis. IEEE Eng Med Biol Mag. 2003;22(1):57–63.

    Article  Google Scholar 

  27. Tzimourta KD, Afrantou T, Ioannidis P, Karatzikou M, Tzallas AT, Giannakeas N, et al. Analysis of electroencephalographic signals complexity regarding Alzheimer’s disease. Comput Electr Eng. 2019;76:198–212.

    Article  Google Scholar 

  28. Şeker M, Özbek Y, Yener G, Özerdem MS. Complexity of EEG dynamics for early diagnosis of Alzheimer’s disease using permutation entropy neuromarker. Comput Methods Programs Biomed. 2021;206: 106116.

    Article  Google Scholar 

  29. Amezquita-Sanchez JP, Mammone N, Morabito FC, Adeli H. A new dispersion entropy and fuzzy logic system methodology for automated classification of dementia stages using electroencephalograms. Clin Neurol Neurosurg. 2021;201: 106446.

    Article  Google Scholar 

  30. Maturana-Candelas A, Gómez C, Poza J, Ruiz-Gómez SJ, Núñez P, Rodríguez M, et al. Analysis of spontaneous EEG activity in Alzheimer’s disease patients by means of multiscale spectral entropy. In International Conference on NeuroRehabilitation, 2018; 579–583.

  31. Santos Toural JE, Montoya Pedrón A, Marañón Reyes EJ. Classification among healthy, mild cognitive impairment and Alzheimer’s disease subjects based on wavelet entropy and relative beta and theta power. Pattern Anal Appl. 2021;24(2):413–22.

    Article  Google Scholar 

  32. Abásolo D, Escudero J, Hornero R, Gómez C, Espino P. Approximate entropy and auto mutual information analysis of the electroencephalogram in Alzheimer’s disease patients. Med Biol Eng Compu. 2008;46:1019–28.

    Article  Google Scholar 

  33. Chai X, Weng X, Zhang Z, Lu Y, Liu G, Niu H. Quantitative EEG in mild cognitive impairment and Alzheimer’s disease by AR-spectral and multi-scale entropy analysis. In World Congress on Medical Physics and Biomedical Engineering, 2019; 2: 159–163.

  34. Staudinger T, Polikar R. Analysis of complexity based EEG features for the diagnosis of Alzheimer’s disease. In 2011 Annual International Conference of the IEEE Engineering in Medicine and Biology Society, 2011; 2033–2036.

  35. Houmani N, Dreyfus G, Vialatte FB. Epoch-based entropy for early screening of Alzheimer’s disease. Int J Neural Syst. 2015;25(08):1550032.

    Article  Google Scholar 

  36. Al-Nuaimi AH, Jammeh E, Sun L, Ifeachor E. Tsallis entropy as a biomarker for detection of Alzheimer’s disease. In 2015 37th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), 2015;4166–4169.

  37. Siuly S, Alçin ÖF, Kabir E, Şengür A, Wang H, Zhang Y, Whittaker F. A new framework for automatic detection of patients with mild cognitive impairment using resting-state EEG signals. IEEE Trans Neural Syst Rehabil Eng. 2020;28(9):1966–76.

    Article  Google Scholar 

  38. Simons S, Abasolo D, Escudero J. Classification of Alzheimer’s disease from quadratic sample entropy of electroencephalogram. Healthc Technol Lett. 2015;2(3):70–3.

    Article  Google Scholar 

  39. Lal U, Chikkankod AV, Longo L. A comparative study on feature extraction techniques for the discrimination of frontotemporal dementia and Alzheimer’s disease with electroencephalography in resting-state adults. Brain Sci. 2024;14(4):335.

    Article  Google Scholar 

  40. Cataldo A, Criscuolo S, De Benedetto E, Masciullo A, Pesola M, Schiavoni R. A novel metric for Alzheimer’s disease detection based on brain complexity analysis via multiscale fuzzy entropy. Bioengineering. 2024;11(4):324.

    Article  Google Scholar 

  41. Wang R, He Q, Shi L, Che Y, Xu H, Song C. Automatic detection of Alzheimer’s disease from EEG signals using an improved AFS–GA hybrid algorithm. Cogn Neurodynamics. 2024. https://doiorg.publicaciones.saludcastillayleon.es/10.1007/s11571-024-10130-z.

    Article  Google Scholar 

  42. Liu J, Lu H, Zhang X, Li X, Wang L, Yin S, Cui D. Which multivariate multi-scale entropy algorithm is more suitable for analyzing the EEG characteristics of mild cognitive impairment? Entropy. 2023;25(3):396.

    Article  Google Scholar 

  43. Sadiq MT, Yu X, Yuan Z, Aziz MZ, Ur Rehman N, Ding W, Xiao G. Motor imagery BCI classification based on multivariate variational mode decomposition. IEEE Trans Emerg Topics Comput Intell. 2022;6(5):1177–89.

    Article  Google Scholar 

  44. Xiao-Feng L, Yue W. Fine-grained permutation entropy as a measure of natural complexity for time series. Chin Phys B. 2009;18(7):2690.

    Article  Google Scholar 

  45. Pincus SM. Approximate entropy as a measure of system complexity. Proc Natl Acad Sci. 1991;88(6):2297–301.

    Article  MathSciNet  Google Scholar 

  46. Richman JS, Moorman JR. Physiological time-series analysis using approximate entropy and sample entropy. Am J Physiol-Heart Circ Physiol. 2000;278(6):2039–49.

    Article  Google Scholar 

  47. Zhang R, Xu P, Chen R, Li F, Guo L, Li P, et al. Predicting inter-session performance of SMR-based brain–computer interface using the spectral entropy of resting-state EEG. Brain Topogr. 2015;28:680–90.

    Article  Google Scholar 

  48. Manis G, Aktaruzzaman MD, Sassi R. Bubble entropy: an entropy almost free of parameters. IEEE Trans Biomed Eng. 2017;64(11):2711–8.

    Article  Google Scholar 

  49. Hsu CF, Wei SY, Huang HP, Hsu L, Chi S, Peng CK. Entropy of entropy: measurement of dynamical complexity for biological systems. Entropy. 2017;19(10):550.

    Article  Google Scholar 

  50. Zunino L, Pérez DG, Garavaglia M, Rosso OA. Wavelet entropy of stochastic processes. Physica A. 2007;379(2):503–12.

    Article  Google Scholar 

  51. Xiang J, Li C, Li H, Cao R, Wang B, Han X, Chen J. The detection of epileptic seizure signals based on fuzzy entropy. J Neurosci Methods. 2015;243:18–25.

    Article  Google Scholar 

  52. Rostaghi M, Azami H. Dispersion entropy: a measure for time-series analysis. IEEE Signal Process Lett. 2016;23(5):610–4.

    Article  Google Scholar 

  53. Yang J, Choudhary GI, Rahardja S, Franti P. Classification of interbeat interval time-series using attention entropy. IEEE Trans Affect Comput. 2020;14(1):321–30.

    Article  Google Scholar 

  54. Goshvarpour A, Goshvarpour A. Diagnosis of epileptic EEG using a lagged Poincare plot in combination with the autocorrelation. SIViP. 2020;14(7):1309–17.

    Article  Google Scholar 

  55. Sharma RR, Varshney P, Pachori RB, Vishvakarma SK. Automated system for epileptic EEG detection using iterative filtering. IEEE Sensors Letters. 2018;2(4):1–4.

    Article  Google Scholar 

  56. Sharif B, Jafari AH. Prediction of epileptic seizures from EEG using analysis of ictal rules on Poincaré plane. Comput Methods Programs Biomed. 2017;145:11–22.

    Article  Google Scholar 

  57. Parsi A, Glavin M, Jones E, Byrne D. Prediction of paroxysmal atrial fibrillation using new heart rate variability features. Comput Biol Med. 2021;133: 104367.

    Article  Google Scholar 

  58. Akbari H, Sadiq MT, Payan M, Esmaili SS, Baghri H, Bagheri H. Depression detection based on geometrical features extracted from SODP shape of EEG signals and binary PSO. Traitement du Signal. 2021;38(1):13–26.

    Article  Google Scholar 

  59. Akbari H, Sadiq MT, Jafari N, Too J, Mikaeilvand N, Cicone A, Serra-Capizzano S. Recognizing seizure using Poincaré plot of EEG signals and graphical features in DWT domain. Bratislava Med J. 2023;124(1):12–24.

    Article  Google Scholar 

  60. Zhou ZH. Machine learning. Springer nature; 2021.

    Book  Google Scholar 

Download references

Acknowledgements

We would like to thank Burcu Oltu and Seda Kibaroğlu for preparing, and recording EEG signals and validating the diseases, respectively.

Funding

This study was conducted with no external funding or financial support.

Author information

Authors and Affiliations

Authors

Contributions

Umut Aslan: investigation, software, and writing—original draft. Mehmet Feyzi Aksahin: conceptualization, methodology, and writing—review and editing. All authors reviewed the manuscript.

Corresponding author

Correspondence to Umut Aslan.

Ethics declarations

Ethics approval and consent to participate

All procedures performed in studies involving human participants were by the ethical standards of the Neurology Department of Baskent University Hospital, set by the Baskent University Institutional Review Board and Ethics Committee.

Competing interests

The authors declare no competing interests.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License, which permits any non-commercial use, sharing, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if you modified the licensed material. You do not have permission under this licence to share adapted material derived from this article or parts of it. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by-nc-nd/4.0/.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Aslan, U., Akşahin, M.F. Detection of Alzheimer and mild cognitive impairment patients by Poincare and Entropy methods based on electroencephalography signals. BioMed Eng OnLine 24, 47 (2025). https://doiorg.publicaciones.saludcastillayleon.es/10.1186/s12938-025-01369-6

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doiorg.publicaciones.saludcastillayleon.es/10.1186/s12938-025-01369-6

Keywords