Accepted Manuscript
A decision support system for automated identification of sleep stages from single-channel EEG signals Ahnaf Rashik Hassan, Abdulhamit Subasi PII: DOI: Reference:
S0950-7051(17)30206-X 10.1016/j.knosys.2017.05.005 KNOSYS 3902
To appear in:
Knowledge-Based Systems
Received date: Revised date: Accepted date:
23 December 2016 5 May 2017 6 May 2017
Please cite this article as: Ahnaf Rashik Hassan, Abdulhamit Subasi, A decision support system for automated identification of sleep stages from single-channel EEG signals, Knowledge-Based Systems (2017), doi: 10.1016/j.knosys.2017.05.005
This is a PDF file of an unedited manuscript that has been accepted for publication. As a service to our customers we are providing this early version of the manuscript. The manuscript will undergo copyediting, typesetting, and review of the resulting proof before it is published in its final form. Please note that during the production process errors may be discovered which could affect the content, and all legal disclaimers that apply to the journal pertain.
ACCEPTED MANUSCRIPT
Highlights
posed.
CR IP T
• A single channel EEG based automated sleep scoring method is pro-
• A novel signal processing technique, namely TQWT is employed for sleep staging. • We introduce bagging to classify sleep stages.
AN US
• Efficacy of the method is confirmed by statistical and graphical analyses.
• The performance of the proposed scheme, compared to the existing
AC
CE
PT
ED
M
ones is promising.
1
ACCEPTED MANUSCRIPT
A decision support system for automated identification of sleep stages from single-channel EEG signals
CR IP T
Ahnaf Rashik Hassana,b,∗, Abdulhamit Subasic a
AN US
Institute of Biomaterials and Biomedical Engineering, University of Toronto, Toronto, ON, Canada b Department of Electrical and Electronic Engineering, Bangladesh University of Engineering and Technology, Dhaka-1000, Bangladesh. c Computer Science Department, College of Engineering, Effat University, Jeddah 21478, Saudi Arabia.
Abstract
A decision support system for automated detection of sleep stages can alleviate the burden of medical professionals of manually annotating a large
M
bulk of data, expedite sleep disorder diagnosis, and benefit research. Moreover, the implementation of a sleep monitoring device that is low-power and
ED
portable requires a reliable and successful sleep stage detection scheme. This article presents a methodology for computer-aided scoring of sleep stages us-
PT
ing singe-channel EEG signals. EEG signal segments are first decomposed into sub-bands using tunable-Q wavelet transform (TQWT). Four statisti-
CE
cal moments are then extracted from the resulting TQWT sub-bands. The proposed scheme exploits bootstrap aggregating (Bagging) for classification.
AC
Efficacy of the feature generation scheme is evaluated using intuitive, statistical, and Fisher criteria analyses. Furthermore, the efficacy of Bagging is evaluated using out-of-bag error analysis. Optimal choices of Bagging ∗
Corresponding Author: Ahnaf Rashik Hassan, Phone: +1-647-3367757, Email:
[email protected]
Preprint submitted to Knowledge-Based Systems
May 8, 2017
ACCEPTED MANUSCRIPT
and TQWT parameters are explicated. The proposed methodology for automated sleep scoring is tested on the benchmark Sleep-EDF database and
CR IP T
DREAMS subjects database. Our methodology achieves 92.43%, 93.69%, 94.36%, 96.55%, and 99.75% accuracy for 2-state to 6-state classification of sleep stages on Sleep-EDF database. Experimental results show that the
algorithmic performance of the automated sleep scoring technique presented herein achieves better performance as compared to the state-of-the-art sleep
AN US
staging algorithms. Besides, the proposed scheme performs equally well for two sleep scoring standards, namely- AASM and R&K. Moreover, the proposed decision support system yields high success rate for identifying sleep
states REM and non-REM 1. It can be anticipated that owing to its use of only one channel of EEG signal, the proposed method will be suitable for
M
device implementation, eliminate the onus of medical professionals of anno-
diagnosis.
ED
tating a large volume of recordings manually, and expedite sleep disorder
Keywords: Sleep, EEG, Classification, Wavelet, Statistical moments.
PT
1. Introduction
CE
Sleep scoring is the first step of sleep research and various sleep disorder diagnosis. Traditionally, sleep scoring is performed by expert sleep techni-
AC
cians based on visual observation of signals. Sleep in humans is composed of cyclic alternations of non-rapid eye movement (NREM) and rapid eye movement (REM) sleep states [1]. NREM phase is yet again classified into four states (S1-S4). Stages S4 and S3 are collectively termed as slow wave sleep owing to the dominance of slow oscillations which emanate from neocortex [2]. 3
ACCEPTED MANUSCRIPT
Again, sleep stages S2 and S1 together are termed as shallow sleep [3]. For visual sleep scoring, nocturnal polysomnography (PSG) is scored by expert
CR IP T
sleep scorers in accordance with Rechtschaffen and Kales’s (R&K) guidelines [4] or a recent set of recommendations by American Academy of Sleep Medicine (AASM) [5]. In R&K guidelines, the recordings are annotated in
any one of the 6 stages- S1-S4, Awake (AWA), and REM. The main alterations of AASM guidelines consist of derivations of EEG, “movement time”
AN US
state abolition, combination of stages S4 and S3 into N3, modification of
some context rules in order to make them simpler, and filter setting guidelines [6] [7].
Sleep state detection by examining the EEG signals visually is problematic for many reasons. Sleep stage assessment by technicians is dependent
M
on human resources, which is expensive. The process is also dependent on rater’s experience and expertise [8]. Moreover, the huge bulk of data that
ED
are inspected per patient make sleep scoring by human experts burdensome and subject to misclassification owing to fatigue [9]. In clinical scenarios,
PT
some cases may demand urgent diagnosis [10] [11]. Manual sleep scoring, due to its time-consuming process, cannot deal with such situations. This
CE
failure, in turn, leads to belated sleep disorder diagnosis. Again, given its consuming nature, human expert based sleep stage screening is unsuitable to handle large data-sets. As a result, large-scale population studies in sleep
AC
research are greatly hindered by this caveat [12] [13]. So, it is evident that there is a great need of an automatic sleep scoring system. Now one might ask- why must the sleep scoring scheme be single-channel based? Developing a wearable, unobtrusive, portable, non-invasive, yet low-power
4
ACCEPTED MANUSCRIPT
device is essential for in-home sleep monitoring. The device must be designed and implemented in such a manner that it will reliably conduct a preliminary
CR IP T
test of the user at home without the requiring a doctor. The development of such a wearable device has gained the interest of sleep research community. A reliable monitoring device of this kind will be instrumental in early diagno-
sis of sleep disordered breathing, such as- sleep apnea and other sleep related disorders. With a view to ensuring portability, convenience of the user, and
AN US
wearability, the automated sleep stage identification scheme employed by the device must be single-channel based. This will also make sure that the battery has an enhanced life and the size of the device is comfortable and convenient for the user. The development and subsequent implementation of such as sleep monitoring system, thus, solely relies on a single-channel based
M
reliable computerized sleep staging scheme.
Various prior studies have attempted to perform sleep stage classifica-
ED
tion [14]. Diykh et al. [15] utilized time domain statistical features and structural graph similarity for feature extraction and K-means in order to
PT
classify sleep stages. Kouchaki et al. [16] used tensor based singular spectrum analysis for computerized identification of sleep-EEG signals. Hassan
CE
et al. [17] put forward a computerized sleep staging scheme employing empirical mode decomposition (EMD) and adaptive boosting. Riaz et al. [18] employed spectral and temporal features extracted in the EMD domain to
AC
devise their computerized sleep scoring scheme. Owing to the use of EMD, both of the aforementioned schemes suffer from mode-mixing problem. Tsinalis et al. [19] implemented a time-frequency analysis based feature extraction and a stacked sparse autoencoders based classification scheme to design
5
AN US
CR IP T
ACCEPTED MANUSCRIPT
Figure 1: A scheme of our computer-aided sleep stage scoring framework.
their system for automatic scoring of sleep-EEG. Peker [20] utilized a combination of dual-tree complex wavelet transform (DT-CWT) and Taguchi
M
based complex-valued neural network for computerized sleep staging from one channel of EEG signals.
ED
In the present study, we present a system for automated scoring of sleep states from single-channel EEG signals. The framework proposed herein is
PT
demonstrated in Fig. 1. The EEG signal segments belonging to various sleep states are first decomposed into sub-bands using a data-driven signal
CE
analysis technique, namely- tunable Q-factor wavelet transform (TQWT). Four statistical moments, namely- mean, variance, skewness, and kurtosis are then extracted from each of the sub-bands. The test and the train ma-
AC
trices are formed afterwards. For feature selection and for determining the statistical significance of the differences of the selected features among all the sleep states, Kruskal-Wallis one-way analysis of variance (Kruskal-Wallis ANOVA) is performed. The features that do not pass Kruskal-Wallis ANOVA
6
ACCEPTED MANUSCRIPT
test are removed and the feature matrices are then fed into the classifier. The classifier used in this work is an eminent ensemble-meta learning based clas-
CR IP T
sification model, namely- bootstrap aggregating (Bagging). Influences and optimal choices of various bagging and TQWT parameters are inspected.
Our experimental findings evince that the algorithmic performance of our
scheme is promising. Algorithmic performance evaluation against the stateof-the-art studies suggest that the proposed sleep scoring scheme performs
AN US
better than or comparable to existing studies in terms of various evaluation metrics. Furthermore, the proposed scheme gives higher detection performance for REM and S1 stages than many of the existing computer-aided sleep scoring schemes.
The organization of the remainder of the article is described below. Sec-
M
tion 2 elucidates the proposed feature extraction framework, investigates its efficacy, describes the statistical analysis that has been performed, and de-
ED
scribes the classifier used in the proposed method. Section 3 expounds the experimentations that have been conducted, presents their results, and expli-
PT
cates their significance. Finally, Section 4 makes some concluding remarks, points to some of the future directions of this work, and brings this article
CE
to conclusion.
AC
2. Materials and methods In this study, Sleep-EDF database available at Physionet and DREAMS
Subjects database have been utilized to conduct the experiments.
7
ACCEPTED MANUSCRIPT
10
15
20
0
5
10
15
20
0
5
10
15
20
0
5
10
15
20
0
5
10
15
0
5
10
15 time (seconds)
25
30
25
30
25
30
25
30
20
25
30
20
25
30
M
50 0 −50
5
CR IP T
50 0 −50
100 0 −100
AWA
REM
50 0 −50
0
AN US
S1
100 0 −100
Amplitude (µV) S4 S3
S2
EEG Epochs of Various Sleep States 100 0 −100
ED
Figure 2: Examples of EEG signal segments of the six sleep states.
2.1. Experimental data
PT
2.1.1. Sleep-EDF data-set
Sleep-EDF is an open-source, benchmark, and extensively utilized database
CE
in sleep scoring literature [21], [22], [23], [24]. Eight Caucasian female and male subjects took part in the study. The subjects’ ages ranged from 21-
AC
35 years. At the time of the study, none of the subjects was not taking any medication. Sleep-EDF has eight recordings. The EEG data have been grouped into two subsets (termed as sc∗ and st∗ ). The first set of EEG data (sc4112e0, sc4102e0, sc4012e0, sc4002e0) have been recorded in a study in 1989. The study subjects were ambulatory healthy. The second set of 8
ACCEPTED MANUSCRIPT
EEG data (st7132j0, st7121j0, st7052j0, st7022j0) have been recorded in a study in 1994. The participants of this study felt slight difficulty in falling
CR IP T
asleep even though all the subjects where health. Data collection has been performed in 24 hours of the subjects’ daily lives. In this subset, nocturnal EEG data have been collected from the four subjects in a hospital with the
help of a miniature telemetry system [25]. For both of the subsets, data of two EEG channels- Fpz-Cz and Pz-Oz have been collected. The sampling
AN US
frequency of the EEG signals is 100 Hz. Prior studies [21], [22], [23], [26] have demonstrated that EEG data from Pz-Oz channel give better algorithmic performance than Fpz-Cz channel. So, data from Pz-Oz channel have been used herein. As per R&K criteria, EEG recordings of both of the subsets have been annotated by an experienced sleep technician on 30s basis.
M
Therefore, the duration of each epoch is 30s or 3000 samples. The epochs have been annotated by sleep technicians as- AWA, REM, S1, S2, S3, S4,
ED
Movement Time, or ‘Unscored’. Fig. 2 shows some EEG signal segments of all the sleep stages. We also consider five cases of classification for both of
PT
the databases used in this study as Table 1 suggests. The six-state case is the R&K scoring itself- AWA, S1-S4, REM. The 5-state case combines S4
CE
and S3 of 6-state as slow wave sleep. The 4-state case combine S2 and S1 of 5-state as Table 1 indicates. The 3-state and 2-state cases consist of- REM, AWA, NREM and Sleep (REM and S1-S4), AWA respectively. Since AASM
AC
guideline combines S3 and S4 into N3 to consider 5 stages of sleep, case I is not considered for AASM scoring. The epoch distribution of each of the states of sleep of this database is shown in Table 2. The total EEG signal segments from Sleep-EDF used in this study is 15,188.
9
ACCEPTED MANUSCRIPT
Table 1: Cases considered for classification in this study.
Number of Classes
Constituent Sleep States
I
6
AWA, REM, S1, S2, S3, S4
II
5
AWA, REM, S1, S2, SWS(S3-S4)
III
4
AWA, REM, S1-S2, SWS(S3-S4)
IV
3
AWA, REM, NREM (S1-S4)
V
2
AWA, Sleep (REM & NREM)
AN US
CR IP T
Case
Table 2: Description of epochs of various sleep states of sleep-EDF database.
S1
S2
S3
S4
REM
Total
8055
604 3621 672
627
1609
15188
M
Epochs
AWA
ED
2.1.2. DREAMS Subjects database
The second database that has been used in this work is DREAMS Subjects database. 20 healthy subjects (16 females, 4 males, 20-65 years old)
PT
who were not taking any medication, participated in this study and 20 nocturnal PSG have been collected from the study participants. The study has
CE
been conducted in the sleep lab of Andr Vsale hospital (Montigny-le-Tilleul, Belgium). Each recording is 7-9 hours in duration. DREAMS Subjects has
AC
168.81 hours of EEG data. The number of EEG segments or epochs extracted from this database is 20,257. A digital 32-channel polygraph (BrainnetTM System of MEDATEC, Brussels, Belgium) has been employed to collect the recordings. DREAMS Subjects database contains one submental EMG chan-
10
ACCEPTED MANUSCRIPT
nel, 3 channels of EEG data (C3-A1 or CZ-A1, O1-A1, and FP1-A1) and 2 channels of EOG data (P18-A1, P8-A1) have been recorded. The data have
CR IP T
been stored in European data format (EDF). The sampling rate of the signals was 200 Hz. Annotations in accordance with both R&K and AASM criteria have been done by an expert sleep technician of the sleep lab. R&K criteria considers sleep states- AWA, REM, S1-S4, ‘sleep stage movement’
and ‘unknow sleep stage’. The unknown sleep stages and movement times
AN US
have been excluded from this study. Again, as per AASM standards, epochs have been annotated in five states- AWA, N1, N2, N3, REM [27]. To carry out the experiments in this study, 50 percent of the Sleep-EDF database has been chosen at random as training data and the other half of the database as testing data. Epochs of each sleep stage were divided
M
equally to ensure that both the training and the testing sets contain epochs of all the sleep stages. The same strategy is adopted for DREAMS Subjects
ED
database as well. In this manner, we have made sure that for each of the two databases, the entire database is used either for training or testing, yet
PT
never both simultaneously. This data distribution decreases the chance of overfitting as well. The process to divide the databases is repeated 20 times.
CE
The mean of the performance metrics of this 20 runs have been presented in this paper.
AC
2.2. Tunable-Q factor wavelet transform Tunable-Q factor wavelet transform (TQWT) is a newly developed signal
decompositions technique that is data-adaptive and flexible. That is why, it has been proven to be an ideal technique to process oscillatory signals [28]. TQWT ensures flexibility by varying its parameters- Q-factor (Q), no. of 11
ACCEPTED MANUSCRIPT
H0j (ω)
LP Scaling αj
(a)
s[n]
H1j (ω)
LP Scaling αj−1
HP Scaling β
dj [n]
AN US
(b)
cj [n]
CR IP T
s[n]
Figure 3: A scheme depicting how an input signal s[n] is decomposed by TQWT up to jth level. (a) the low-pass sub-band signal cj [n] and (b) the high-pass sub-band signal dj [n].
decomposition levels (J), and over-sampling rate (r). Q controls the wavelet’s
ED
M
oscillation. The mathematical formulation of Q is given below.
CE
PT
where
Q=
fc (j) BW (j)
fc (j) = αj
2−β 4α
1 BW (j) = βαj−1 π 2
(1)
(2)
(3)
and j = 1, 2, ..., J. For decomposition of J levels, TQWT converts an
AC
input signal s[n] into J + 1 sub-bands. For this conversion, a filer bank with two channels is used iteratively. TQWT applies the filter bank to the lowpass sub-band. Input s[n] is decomposed into c0 [n] and d1 [n] in each of the stages. Here, c0 [n] denotes the low-pass sub-band with sampling frequency
12
ACCEPTED MANUSCRIPT
of αfs . d1 [n] indicates the high-pass sub-bands with sampling frequency of βfs . The scaling factors are denoted by α and β and fs is the sampling rate
CR IP T
of s[n]. Low-pass filter H0 (ω) with low-pass scaling α is applied to generate c0 [n]. d1 [n] is created by H1 (ω) and high-pass scaling β. To make sure that
TQWT has perfect reconstruction and no redundancy, the values of α and
β must be: 0 < α < 1, 0 < β ≤ 1 and α + β > 1. For the implementation of TQWT, we have used the toolbox provided by Selesnick [28].
AN US
Fig. 3 presents a schematic outline of how a signal s[n] is processed by TQWT up to jth level to produce cj [n] and dj [n]. H0j (ω) is the frequency response of the low-pass sub-bad and H1j (ω) is the frequency response of the high-pass sub-band created after j-levels. Their mathematical formulations are given below.
PT
ED
0,
M
(J)
H0 (ω) =
Q J−1 H0 (ω/αm ), if |ω| ≤ αJ π m=0
(J)
CE
H1 (ω) =
AC
where
if α π ≤ |ω| ≤ π
Q m H1 (ω/αJ−1 ) J−2 m=0 H0 (ω/α ), if(1 − β)αJ−1 π ≤ |ω| ≤ αJ−1 π 0, for otherω ∈ [−π, π]
ω + (β − 1)π ) α+β−1 απ − ω H1 (ω) = θ( ) α+β−1
H0 (ω) = θ(
(4)
J
(5)
(6) (7)
θ(ω) denotes Daubechies filter’s frequency response whose mathematical 13
ACCEPTED MANUSCRIPT
formulation is given below. (8)
CR IP T
p θ(ω) = 0.5(1 + cos(ω)) 2 − cos(ω), |ω| < π TQWT parameters r and Q are related to α and β as: r=
β , 1−α
Q=
2−β β
2.3. Advantages of TQWT
(9)
AN US
The use of TQWT in the proposed scheme is motivated by various advantages of TQWT over conventional signal processing techniques such as empirical mode decomposition (EMD) and Fourier transform. • The Q-factor of the wavelet must be low as the wavelet transform at-
M
tempts to process less oscillatory signals. On the contrary, the Q-factor of the wavelet must be high as the wavelet transform attempts to pro-
ED
cess highly oscillatory (e.g. EEG) signals. Nevertheless, except for the continuous wavelet transform, most wavelet transforms cannot tune the Q-factor [28]. TQWT eradicates this limitation by tuning its param-
PT
eter Q. Therefore, it has been successful in analysis and classification
CE
of EEG.
• Feature generation schemes that use Fourier transform based features,
AC
assume that the signal is stationary and linear even though EEG is highly non-stationary and nonlinear. In spite of their data-adaptive nature, EMD and almost all of its variations are affected by modemixing. TQWT, on the contrary, does not have this caveat. Therefore, it is more suitable for EEG data analysis. 14
ACCEPTED MANUSCRIPT
• TQWT has been proven to be instrumental in analyzing various biosignals successfully of late [29].
CR IP T
• The transfer functions of TQWT filters are rational, contributing to its less computational burden and efficiency. This also makes the frequency domain specification easier [28].
• The advantage of TQWT over commonly used time-frequency based
AN US
transforms is that TQWT is highly flexible. In other words, TQWT can alter its Q-factor to characterize the morphology of the signal in
the frequency domain in a better manner. Other time-frequency based transforms can not offer this flexibility.
• Last but not the least, wavelet transform’s perfect reconstruction prop-
M
erty is inherited by TQWT.
ED
The advantages mentioned above establish TQWT as a prominent signal decomposition scheme of analyzing oscillatory EEG signals and motivate its use in our framework. Histograms of TQWT sub-band of the six sleep phases
PT
in Fig. 4 further corroborate with the appealing characteristics of TQWT. It is to be noted in Fig. 4 that the shapes of histograms of the EEG signals are
CE
different from each other. For instance, histograms corresponding to S1 and S2 are somewhat skewed to the right as opposed to the histograms of other
AC
sleep stages. Also, the dispersion of the histograms are different from each other. Another key difference among the histograms in Fig. 4 is that the ranges of amplitudes of EEG signals in different sleep stages are different as well. The histograms shapes are different for the six phases of sleep in Fig. 4, conforming yet again the efficacy of TQWT for analyzing sleep-EEG. 15
ACCEPTED MANUSCRIPT
S2 State
40
40
40
30 20
Occurrence
50
30 20 10
0 −40
−20
0 Voltage (µV) S4 State
20
−50
0 Voltage (µV) AWA State
50
0 −200
100
60
50
50
40
40
20 10
Occurrence
60
40 Occurrence
50
30
30 20 10
−100
0 Voltage (µV)
100
200
0 −200
−100
0 Voltage (µV) REM State
100
200
30 20 10
−100
0 100 Voltage (µV)
200
300
0 −100
−50
AN US
0 −200
20 10
0 −100
40
30
CR IP T
10
Occurrence
S3 State
50
Occurrence
Occurrence
S1 State 50
0 Voltage (µV)
50
100
Figure 4: TQWT sub-bands’ histograms of the sleep states. Significant differences of histogram shapes are observed among the six stages of sleep. Q = 2, r = 3, J = 15 in this example.
M
2.4. Statistical moment based features
Once we have obtained the TQWT sub-bands, the next step is to extract
ED
four statistical moments to be used as features from each of the sub-bands as Fig. 1 evinces. The use of statistical moments as features in this study
PT
is motivated by their successful and widespread use in physiological signal analysis [30]. As these measures signify the asymmetry, peakedness, and dis-
CE
persion of a data distribution, one can anticipate that statistical features will capture the underlying differences of the TQWT sub-bands corresponding to
AC
the six classes illustrated in Fig. 4. This hypothesis is validated by both the good discrimination power and classification performance of the features as we shall see in subsequent sections of this article. For a TQWT sub-band with length ζ, T = τ1 , τ2 , τ3 , .., τN mean (θ), variance (σ 2 ), skewness (ι), and kurtosis (ν) can be formulated as shown in Table 3. 16
ACCEPTED MANUSCRIPT
Table 3: Description of the four features used in the proposed scheme.
Mathematical Formulation
Mean (θ)
θ= σ2 =
Skewness (ι)
ι=
1 ζ
ν=
1 ζ
i=1 τi
2 τi − θ 3
i=1
PN
i=1
PN
i=1
τi −θ σ
τi −θ σ
4
AN US
Kurtosis (ν)
PN
PN
1 ζ
Variance (σ 2 )
1 ζ
CR IP T
Feature Name
2.5. Separability evaluation
The efficacy of statistical moments extracted from TQWT sub-bands is yet again demonstrated herein by separability analysis. Our feature genera-
M
tion scheme’s effectiveness is now evaluated using Fisher criteria [31]. Unlike most of the separability metrics, Gaussian assumption of the data in consid-
ED
eration is not required for Fisher criteria [31]. The two Fisher criteria are
PT
expressed mathematically as follows. J1 =
tr(Sm ) tr(Sw )
J2 = tr(Sw−1 Sm )
(10) (11)
CE
where Sw and Sm evince the within-class and the between-class scatter
matrices respectively and tr(S) denotes the trace of S. The larger the values
AC
of J2 and J1 are, the features in consideration are more separable in feature space. Table 4 compares J1 and J2 values of our method with two approaches presented in [32]. It is evident from Table 4 that the J1 and J2 values of
our scheme are higher. This indicates that statistical moments from TQWT sub-bands give high separability among the six sleep phases. 17
ACCEPTED MANUSCRIPT
Table 4: Comparison of J1 and J2 of various schemes on Sleep-EDF.
J1
Power spectrum density features [32] STFT and RVM [32] Statistical features and CEEMDAN [30]
0.69
1.82
0.74
4.57
0.88
56.757
0.68
66.4
AN US
Spectral features from sleep-EEG signals [33] Proposed Method
2.6. Feature selection
J2
CR IP T
Methods
1.3147 125.5887
M
With a view to ascertaining if the values of the four statistical features differ significantly in the six sleep stages (AWA, S1-S4, and REM) and if
ED
the discrimination ability of the four statistical features extracted from the TQWT domain are significant statistically, a Kruskal-Wallis one-way analy-
PT
sis of variance is carried out. An advantage of Kruskal-Wallis test over other hypothesis testing schemes is that it does not require the underlying data
CE
distribution to be normal [34]. Kruskal-Wallis ANOVA is performed at a confidence level of 95%. Consequently, a difference is regarded to have sig-
AC
nificance if p < α(= 0.05) is found. The result of Kruskal-Wallis ANOVA is presented in Table 5. We can see that features computed from some of the sub-bands do not pass the hypothesis test. Before classification using the test and the train matrices, these features are removed. Additionally, we can see from Table 5 that of the features that pass the test, almost all of them
18
ACCEPTED MANUSCRIPT
Table 5: p-Values of θ, σ 2 , ι, and ν across various TQWT sub-bands for case I. Statistically insignificant features are shown in bold.
PT
ι 0.002586 0.003781 0.000514 0.0157 0.1539 0.002926 0.009327 0.002203 0.00999 0.003088 0.005488 0.1265 0.6956 0.008618 0.001954 0.002598 0.007907 0.008866 0.003938 0.008379 0.0085 0.6936 0.006309 0.009433
ν 1.3926e-277 8.7060e-291 8.3727e-243 7.3642e-158 5.2075e-57 2.3317e-277 3.7052e-173 1.6636e-93 4.9284e-94 1.3895e-124 4.5837e-144 2.1401e-134 2.4296e-110 1.4980e-92 6.2755e-69 1.4292e-64 1.4344e-34 7.4337e-40 2.9448e-46 3.0988e-30 6.3235e-29 6.5229e-36 2.1707e-17 1.9109e-07
CR IP T
σ2 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
M
AN US
θ 0.6762 0.002916 0.000821 0.625 0.000585 0.008186 0.00824 0.005205 0.000957 0.0009891 0.008398 0.004177 0.006163 0.002103 0.00026 0.000953 0.001756 0.007923 0.004003 0.00344 0.9744 0.00396 0.00723 0.000564
ED
Sub-band 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24
CE
give lower p-values evincing good inter-class variation of the features. Thus, the results of Kruskal-Wallis ANOVA confirm the inferences drawn from Fig.
AC
4 and from Fisher criteria evaluation that the four statistical moments can characterize the sleep stages successfully and the proposed feature extraction scheme is efficacious.
19
ACCEPTED MANUSCRIPT
Algorithm 1: Bagging Input: N number of training examples S = [(xi , yi )] of γ classes having labels yi ∈ ω, ω = {ω1 , ω2 , ...., ωγ } and i = 1, 2, 3, ....., N . Base
CR IP T
learner or weak classifier or learner WeakLearn. A fraction φ to form
bootstrapped training data-set. T , an integer that indicates the total iterations.
for t = 1, 2, ...., T do 1. Draw a bootstrapped replica St by taking φ percent of S at
AN US
random.
2. Call WeakLearn with St and get the hypothesis (classifier) νt . νt to the ensemble η. end
Test: For a test instance, x with unknown label:
M
1. Evaluate η = {ν1 , ν2 , ..., νT } on x. 1,
ED
2. vt,j =
0,
if νt picks class ωj Otherwise
CE
PT
3. Find the total vote obtained by each class: Vj =
T X
vt,j
t=1
where j = 1, 2, .., γ
AC
4. Select the class that has obtained the highest number of votes as bagging’s output.
20
ACCEPTED MANUSCRIPT
2.7. Bootstrap aggregating Bootstrap aggregating (Bagging) is an ensemble-meta algorithm first pro-
CR IP T
posed in [35] which creates and combines multiple classification models to solve a particular classification problem. It is one of the simplest and most
intuitive to implement ensemble learning based classifier, even though it yields strikingly good performance [36]. Bagging is depicted Algorithm 1.
Bagging achieves diversity by employing original training data-set’s boot-
AN US
strapped replicas. First, subsets are taken at random with replacement from the original training data-set. Afterwards, each subset St drawn as such is utilized to train a different weak classifier, generally of the same type. In this
way, a hypothesis νt is obtained each time. This weak classifier or predictor is represented as WeakLearn in Algorithm 1. In this article, we use deci-
M
sion trees as our weak classifier or learning algorithm. Individually obtained hypotheses are combined afterwards by means of majority voting. For any
ED
given unlabeled test instance, the ensemble’s decision is the class chosen by most weak classifiers. The number of iterations T and WeakLearn parame-
PT
ters (e.g. number of leaves for decision tree) affect Bagging’s performance as Algorithm 1 evinces. It is to be noted that T is equal to the total number of
CE
decision trees (WeakLearn) since one tree is trained in each iteration. The optimal choice of T is important. If T is too small or too large, it will result
AC
in poor performance or computational inexpensiveness of the algorithm respectively [33]. We shall expound how T and number of leaves are selected
in the next section.
21
ACCEPTED MANUSCRIPT
OOB classification error vs. number of grown trees for case I to V 0.4 Case I Case II Case III Case IV Case V
CR IP T
0.3 0.25 0.2 0.15
AN US
Out−of−bag classification error
0.35
0.1 0.05
0
50
100 150 200 Number of decision trees
250
300
M
0
ED
Figure 5: Variation of OOB error with the number of decision trees.
3. Results and discussions
PT
Here we explicate the details and the outcomes of our experiments and expound some of the significance of the results.
CE
With a view to finding the optimal decision trees required to obtain the
best algorithmic performance and to examining how the accuracy value varies
AC
with the no. of trees, we find out the no. of decision trees which gives minimum out-of-bag (OOB) error. This process is repeated every time for cases I to V. The resulting values are used to classify the unlabeled test instances in each case. Fig. 5 depicts the variation of OOB error as we vary no. of trees from 1 to 300 for cases I to V. Fig. 5 manifests that after about 22
ACCEPTED MANUSCRIPT
LOOCV error variation with number of decision trees 18 Case I Case II Case III Case IV Case V
CR IP T
16
12 10 8
AN US
LOOCV error (%)
14
6 4
0
50
100 150 200 Number of decision trees
250
300
M
2
Figure 6: Variation of LOOCV error with number of decision trees reveals the prediction
ED
power of the classifier in all the cases.
PT
20 decision trees, the OOB error almost becomes constant. However, to make sure that the algorithm yields the best performance, no. of trees for which
CE
OOB error is minimum is noted every time for cases I to V. Afterwards, when classifying the test examples of each case, these values are used. For cases I to V, the optimal no. of decision trees are- 285, 265, 153, 295, and 188
AC
respectively.
Fig. 5 also manifests the prediction power of bagging for the proposed
feature extraction scheme. In fact, prior studies have empirically demonstrated that OOB estimation is tantamount to using a test data-set having
23
ACCEPTED MANUSCRIPT
Variation of accuracy values (%) with the variation of r for cases I−V. 98
CR IP T
94 92 90 88
Case V Case IV Case III Case II Case I
86 84
4
6
8
10
AN US
Accuracy in percentage
96
12
14
16
18
20
M
Value of r
Figure 7: Accuracy variation with r.
ED
equal size of a train data-set [35]. As OOB error estimates the true ensemble error unbiasedly, cross-validation or a separate test data-set to assess the
PT
prediction ability of the classification model is not needed anymore. In classification cases where insufficient data are available, this approach for analyzing
CE
the predictive performance is really helpful. The low OOB error of Fig. 5 also manifests good predictive capability of bagging. Furthermore, after 20
AC
trees, OOB error variation becomes trivial, indicating the robustness of our method. To further validate the prediction power of the classification model, we have conducted leave-one-out cross-validation (LOOCV) analysis. Fig. 6 shows that the proposed decision support system yields low LOOCV error for a wide range of number of decision trees for all the five cases of interest. 24
ACCEPTED MANUSCRIPT
Variation of accuracy values (%) with the variation of Q for cases I−V. 98 96
CR IP T Case V Case IV Case III Case II Case I
92 90 88 86 84 82 80 78
5
10
15
20
AN US
Accuracy in percentage
94
25
30
35
40
45
50
M
Value of Q−factor, Q
Figure 8: Accuracy variation with Q-factor Q.
ED
Thus, LOOCV further validates the prediction capability of the algorithm. The foregoing findings are later validated and confirmed by bagging’s high
PT
detection accuracy of classifying test data-set. Prior works on TQWT indicate that r ≥ 3 to circumvent excessive un-
CE
desirable wavelet ringing [28] and higher detection accuracy is obtained if r is lower [37]. The experimental findings of this study confirm this as well.
AC
Fig. 7 shows how the variation of the value of TQWT parameter r influences the algorithmic performance of the proposed sleep scoring framework. It is clear from Fig. 7 that the highest detection accuracy is obtained when the values of r are set to- 4, 5, 4, 4, and 5 for cases I to V. So, these values of r are used in this study. Again, for Q, the experiments indicate that the 25
ACCEPTED MANUSCRIPT
Variation of accuracy values (%) with the number of decomposition levels, J for cases I−V. 98
CR IP T
94 92 90 88
Case V Case IV Case III Case II Case I
AN US
Accuracy in percentage
96
86 84 82
5
10
15
20
25
30
M
Value of number of decomposition levels, J
Figure 9: Accuracy variation with number of decomposition levels J.
ED
highest detection accuracy is obtained when Q is set to 2, 2, 2, 3, and 1 respectively for cases I to V. This is confirmed by many experimental obser-
PT
vations wherein the values of the other two parameters J and r have been varied. Fig. 8 serves as a pictorial demonstration of this fact. The detection
CE
accuracy vs. J does not demonstrate a decreasing trend like Q and r as Fig. 9 evinces. Here the value of J is varied in the range of 1 to 30 in step of one
AC
setting the optimal values of r and Q in each of the five cases to find out the values of J for which detection accuracy is maximum. The highest detection accuracy is obtained when the value of J is set to 23, 21, 22, 14, and 15.
The parameter value choices have been confirmed and validated from many experimentations. Thus, it can be concluded that the aforementioned values 26
ACCEPTED MANUSCRIPT
Table 6: Confusion matrix for case I.
Predicted Class S3
S1
113
43
0
S2
3
1682
23
S3
0
141
116
S4
0
5
27
AWA
27
4
0
REM
20
84
S4
AWA
REM
Sen
CR IP T
S2
0
56
90
37.42%
8
58
37
92.93%
53
26
0
34.52%
264
18
0
84.35%
0
3971
26
98.61%
1
38
662
82.34%
AN US
Actual Class
S1
0
Table 7: Confusion matrix for case II.
Predicted Class
S3-S4
AWA
REM
Sen
43
0
53
89
38.74%
S2
5
1641
54
70
41
90.66%
S3-S4
0
104
513
33
0
79.04%
AWA
28
6
0
3971
23
98.61%
REM
24
84
1
41
655
81.47%
CE
PT
Actual Class
117
ED
S1
S2
M
S1
are the optimal values of TQWT parameters.
AC
The confusion matrices presented in Tables 6-10 give further insights on
the algorithmic performance of our scheme for cases I to V. It is clear from the matrices that in general, as classes increase, the detection accuracy of each sleep stage decreases. This is also reflected on the overall accuracy
27
ACCEPTED MANUSCRIPT
Table 8: Confusion matrix for case III.
Predicted Class AWA
REM
Sen
S1-S2
1879
60
85
89
88.97%
S3-S4
101
515
34
0
79.35%
AWA
63
0
3945
20
97.96%
REM
157
1
32
615
76.49%
CR IP T
S3-S4
AN US
Actual Class
S1-S2
Table 9: Confusion matrix for case IV.
Predicted Class
Actual Class
AWA
REM
Sen
2592
88
83
93.88%
65
3940
23
97.84%
163
41
601
74.75%
ED
REM
AWA
M
S1-S4
S1-S4
CE
PT
Table 10: Confusion matrix for case V.
S1-S4, REM
AWA
Sen
S1-S4, REM
3467
101
97.25%
AWA
113
3915
97.22%
AC
Actual Class
Actual Class
values for cases I to V. The proposed sleep scoring framework consistently yields high detection accuracy for all the sleep stages except for sleep stage S3 in case I as Table 6 indicates. One of the main reasons of poor per-
28
ACCEPTED MANUSCRIPT
Table 11: S1 detection accuracy of various methods for case I on Sleep-EDF data-set. The highest accuracy value is highlighted in bold.
S1
Doroshenkov et al. [38]
4.84%
Zhu et al. [21]
15.8%
Liang et al. [22]
18.75%
Liang et al. [39]
30%
AN US
CR IP T
Authors
33.70%
Ronzhina et al. [23]
36.17%
Hsu et al. [41]
36.70%
Proposed Method
37.42%
M
Vural et al. [40]
Table 12: S1 detection accuracy of various methods for case II on Sleep-EDF data-set.
AC
CE
PT
ED
The highest accuracy value is highlighted in bold.
Authors
S1
Zhu et al. [21]
15.8%
Liang et al. [22]
18.75%
Liang et al. [39]
30%
Hsu et al. [41]
36.70%
Proposed Method 38.74%
formance of computerized sleep staging algorithms is the resemblance of S1 state’s EEG signals with those of REM’s. In fact, S1 and REM are almost
29
ACCEPTED MANUSCRIPT
Table 13: Comparison of performance of various works on Sleep-EDF. Only the highest accuracy value (%) of each method is reported.
Fuzzy logic based iterative method [26]
CR IP T
Case ICase IICase IIICase IVCase V -
71.2
74.5
88.3
95.4
-
87.20
-
-
-
-
83.60
-
-
-
-
77.98
-
-
-
76.70
-
81.42
88.97
96.90
Temporal features and hidden Markov model [38] 61.08
-
-
-
-
-
-
-
-
Moment based features and adaptive boosting [42] 80.34 82.03
82.83
-
-
M
Methods
Frequency bands’ energy, neural network [41] Multiscale entropy, LDA [22]
AN US
Temporal features and hierarchical decision tree [39] PSD of single-channel EEG and ANN [23]
Hybrid features, Karhunen-Love transform [40]
69.98
85.57 86.53
87.49
89.77
95.05
Graph theory based features, SVM [21]
87.50 88.90
89.30
92.60
97.90
ED
Statistical features and bagging [33]
92.43 93.69 94.36 96.55 99.75
PT
Proposed Method
indistinguishable by visual inspection [21]. This is why, most S1 epochs are
CE
misclassified as REM by existing automatic sleep scoring methods. This is also evident from case I confusion matrix in Table 6 and case II confusion
AC
matrix in Table 7. An advantage of the algorithm propounded herein is that it classifies S1 state better than many of the prior studies. As we can see from the case I confusion matrix in Table 6, our method correctly classifies 37.42% S1 EEG epochs. The same scenario is seen in Table 7 as our method correctly classifies 38.74% S1 EEG epochs in case II. Moreover, in compar30
ACCEPTED MANUSCRIPT
ison with the extant state-of-the-art algorithms’ S1 detection performance, the proposed method’s S1 stage identification accuracy is significantly better
CR IP T
as Tables 11 and 12 suggest. Again, for sleep stage classification, REM phase is very important. REM stage consists of about 5-20% of an adult human’s
nocturnal sleep. Furthermore, detection of REM is a precondition for diagnosing many sleep ailments including narcolepsy, RBD etc. [43]. During
REM, muscle activity is observed to diagnose REM. It is a marker of many
AN US
neurological disorders such as Parkinsons disease. So, REM detection bears
particular significance in the context of sleep data analysis. It can be seen from Tables 6-9 that the proposed algorithm gives 82.34%, 81.47%, 76.49%, and 74.75% accuracy for REM detection for cases I to IV. In comparison, the methods put forward in [21] and [40] give 76.20% and 79.70% REM de-
M
tection accuracy respectively for case I as opposed to our method’s 82.34% REM detection accuracy in case I. So, REM phase detection success of our
ED
algorithm is comparable to or better than the existing studies. To get further insight on the efficacy of the proposed automated sleep
PT
scoring scheme, the performance of our method is compared with those of the existing works. For fair comparison, classification results obtained from
CE
Sleep-EDF are considered. Table 13 lists the accuracy values of various studies that use Sleep-EDF. All the accuracy values reported in this section and in Table 13 are the best detection accuracy for a particular method. It
AC
is clear from Table 13 that the method proposed in this work outperforms others in all the cases. To evaluate the effectiveness of the scheme propounded herein even fur-
ther, we compute Cohen’s kappa coefficient for each of the five cases consid-
31
ACCEPTED MANUSCRIPT
Table 14: Kappa coefficients of the proposed method against various published methods on Sleep-EDF data-set.
-
0.7968
0.81
II
0.7452
0.8
0.83
III
-
0.8125
0.83
IV
-
0.84
0.87
V
-
0.9
0.96
AN US
I
CR IP T
Case Liang et al. [22] Hassan et al. [30] Zhu et al. [21] Proposed Method 0.836
0.8543 0.863
0.8933 0.9435
ered in this work. Cohens kappa measures the agreement among annotators for categorial items statistically. It has been considered as a more robust measurement of performance than percent agreement since kappa considers the
M
coincidentally occurring agreements. Kappa coefficient ranging from 0.81-1 is regarded as almost perfect agreement, 0.61-0.80 as substantial agreement,
ED
0.41-0.60 as moderate agreement, 0.21-0.40 as fair agreement, 0-0.20 as slight agreement, and less than zero evinces no agreement. Table 14 evinces that
PT
the proposed scheme gives almost perfect agreement in all the cases of interest. Also, it gives higher accuracy values in comparison with prior works in
CE
all the cases except in case V. We now present the detection performance of our scheme on DREAMS
AC
Subjects. Table 15 lists the kappa coefficient and detection accuracy of our algorithm for AASM and R&K standards of sleep scoring. Even though the detection performance of our scheme is promising for all the cases, for case II in both of the scoring standards, accuracy values of case II are on the lower side. However, it can be observed that the proposed algorithm yields consis32
ACCEPTED MANUSCRIPT
Table 15: Algorithmic performance of the proposed method on DREAMS Subjects dataset.
R&K
76.39
74.39
Accuracy (%)
AASM
-
78.95
Cohen’s Kappa
R&K
0.785
0.78
Cohen’s Kappa
AASM
-
0.82
81.87
88.98
99.02
83.78
89.04
98.75
0.83
0.882
0.95
0.865
0.91
0.985
AN US
Accuracy (%)
CR IP T
Evaluation Metric Scoring Scheme Case I Case II Case III Case IV Case V
tently high detection accuracy in general in each of the five cases. Moreover, Table 15 reveals that automated sleep scoring gets more and more difficult for the algorithm with the increase of classes. This result supports the findings of
M
prior studies in Table 13. Furthermore, our scheme’s detection performance is quite similar in AASM and R&K scoring as Table 15 manifests. So, our
ED
sleep scoring system can be employed equally for AASM and R&K scoring scheme without any algorithmic changes or tuning of parameters, which is a big advantage of the system.
PT
Even though the suitability of TQWT for sleep data analysis has been demonstrated in [44], the proposed method, due to the use of the simple
CE
and computationally inexpensive statistical features, is particularly suitable for device implementation. apart from application in the clinical scenarios
AC
for assisting clinicians, one of the goals of developing an automated decision support system for sleep scoring is to implement the algorithm in a wearable and portable sleep monitoring device for in-home care. This necessitates the sleep scoring algorithm to be computationally as inexpensive as possible. The method proposed in this work not only gives high accuracy values, but it is 33
ACCEPTED MANUSCRIPT
computationally less expensive. This is owing to the fact that the proposed scheme involves computation of four simple statistical features, whereas the
CR IP T
Fourier domain based feature extraction scheme increases the computational complexity and expensiveness of the methodology presented in [44]. Thus,
the method presented in this work is more suitable for device implementation than the one presented in [44].
AN US
4. Conclusions and future directions
In this work, a methodology for computer-aided sleep scoring is presented. Sleep-EEG signal segments are first decomposed using TQWT and four statistical moments are extracted and employed as features. Bootstrap aggregating is employed for classification. Our experimental findings manifest that
M
statistical moment based features can capture the differences of various sleep states successfully. The influences of various Bagging and TQWT param-
ED
eters have been demonstrated. The algorithmic performance of our sleep scoring system as opposed to the existing ones is promising. Further, the
PT
proposed scheme yields good REM and S1 detection accuracy. Moreover, the proposed scheme equally performs well for R&K and AASM sleep scor-
CE
ing standards. In the future, we shall focus on hardware implementation of the proposed algorithm. A combination of various ensemble learning based
AC
classification models for the proposed feature extraction framework can be an interesting avenue of further research. Deep learning can also be employed to further meliorate classification performance in all the cases of interest. Future works can also employ time-frequency based methods [45] [46] such as multi-directional kernel methods [47] [9] and wrapper based feature selection 34
ACCEPTED MANUSCRIPT
techniques for computerized sleep scoring. To conclude, it can be said that the proposed scheme is efficient and efficacious and will alleviate the burden
References
CR IP T
of sleep technicians and expedite sleep disorder research and diagnosis.
[1] J. A. Hobson, Sleep is of the brain, by the brain and for the brain, Nature 437 (7063) (2005) 1254–1256.
AN US
[2] L. Marshall, H. Helgadottir, M. Molle, J. Born, Boosting slow oscillations during sleep potentiates memory, Nature 444 (7119) (2006) 610– 613.
[3] H. tieng Wu, R. Talmon, Y.-L. Lo, Assess sleep stage by modern signal
M
processing techniques, Biomedical Engineering, IEEE Transactions on
ED
62 (4) (2015) 1159–1168.
[4] A. Rechtschaffen, A. Kales, Manual of standardized terminology, techniques and scoring systems for sleep stages of human subjects, U. G. P.
PT
Office, Washington DC Public Health Service.
CE
[5] A. L. C. C. Iber, S. Ancoli-Israel, S. F. Quan, Westchester, usa: American academy of sleep medicine, The AASM Manual for the Scoring of
AC
Sleep and Associated Events: Rules, Terminology and Technical Specification.
[6] K. Samiee, P. Kovcs, M. Gabbouj, Epileptic seizure detection in longterm EEG records using sparse rational decomposition and local gabor binary patterns feature extraction, Knowledge-Based Systems (2016) –. 35
ACCEPTED MANUSCRIPT
[7] U. R. Acharya, H. Fujita, V. K. Sudarshan, S. Bhat, J. E. Koh, Application of entropies for automated diagnosis of epilepsy using EEG signals:
CR IP T
A review, Knowledge-Based Systems 88 (2015) 85 – 96. [8] S. B. Akben, A. Alkan, Visual interpretation of biomedical time series using parzen window-based density-amplitude domain transformation, PLOS ONE 11 (9) (2016) 1–13.
AN US
[9] B. Boashash, S. Ouelha, Automatic signal abnormality detection using
time-frequency features and machine learning: A newborn EEG seizure case study, Knowledge-Based Systems 106 (2016) 38 – 50. [10] S. Kara, F. Dirgenali, A system to diagnose atherosclerosis via wavelet transforms, principal component analysis and artificial neural networks,
M
Expert Systems with Applications 32 (2) (2007) 632 – 640.
ED
[11] B. Karlik, S. Aydin, An improved approach to the solution of inverse kinematics problems for robot manipulators, Engineering Applications
PT
of Artificial Intelligence 13 (2) (2000) 159 – 164. [12] H. G¨ ur¨ uler, A novel diagnosis system for parkinson’s disease using
CE
complex-valued artificial neural network with k-means clustering feature weighting method, Neural Computing and Applications (2016) 1–10.
AC
[13] H. M. Ertunc, H. Ocak, C. Aliustaoglu, Ann- and anfis-based multistaged decision algorithm for the detection and diagnosis of bearing faults, Neural Computing and Applications 22 (1) (2013) 435–446.
[14] A. Subasi, Automatic recognition of alertness level from EEG by using 36
ACCEPTED MANUSCRIPT
neural network and wavelet coefficients, Expert Systems with Applications 28 (4) (2005) 701 – 711.
CR IP T
[15] M. Diykh, Y. Li, P. Wen, Eeg sleep stages classification based on time domain features and structural graph similarity, IEEE Transactions on
Neural Systems and Rehabilitation Engineering PP (99) (2016) 1–1. doi:10.1109/TNSRE.2016.2552539.
AN US
[16] S. Kouchaki, S. Sanei, E. L. Arbon, D. J. Dijk, Tensor based singular
spectrum analysis for automatic scoring of sleep eeg, IEEE Transactions on Neural Systems and Rehabilitation Engineering 23 (1) (2015) 1–9. [17] A. R. Hassan, M. I. H. Bhuiyan, Automatic sleep scoring using statistical features in the emd domain and ensemble methods, Biocybernetics and
M
Biomedical Engineering 36 (1) (2016) 248 – 255.
ED
[18] F. Riaz, A. Hassan, S. Rehman, I. K. Niazi, K. Dremstrup, Emd-based temporal and spectral features for the classification of eeg signals using
PT
supervised learning, IEEE Transactions on Neural Systems and Rehabilitation Engineering 24 (1) (2016) 28–35.
CE
[19] O. Tsinalis, P. M. Matthews, Y. Guo, Automatic sleep stage scoring using time-frequency analysis and stacked sparse autoencoders, Annals
AC
of Biomedical Engineering 44 (5) (2016) 1587–1597.
[20] M. Peker, An efficient sleep scoring system based on eeg signal using complex-valued machine learning algorithms, Neurocomputing (2016) –.
37
ACCEPTED MANUSCRIPT
[21] G. Zhu, Y. Li, P. . Wen, Analysis and classification of sleep stages based on difference visibility graphs from a single-channel eeg signal, IEEE
CR IP T
Journal of Biomedical and Health Informatics 18 (6) (2014) 1813–1821. [22] S.-F. Liang, C.-E. Kuo, Y.-H. Hu, Y.-H. Pan, Y.-H. Wang, Automatic stage scoring of single-channel sleep eeg by using multiscale entropy and autoregressive models, Instrumentation and Measurement, IEEE
AN US
Transactions on 61 (6) (2012) 1649–1657.
[23] M. Ronzhina, O. Janouek, J. Kolrov, M. Novkov, P. Honzk, I. Provaznk, Sleep scoring using artificial neural networks, Sleep Medicine Reviews 16 (3) (2012) 251 – 263.
[24] A. R. Hassan, M. I. H. Bhuiyan, Automatic sleep stage classification,
M
in: Electrical Information and Communication Technology (EICT), 2013
ED
International Conference on, 2015, pp. 1–6. [25] B. Kemp, A. Zwinderman, B. Tuk, H. Kamphuisen, J. Oberye, Analysis of a sleep-dependent neuronal feedback loop: the slow-wave microconti-
PT
nuity of the eeg, Biomedical Engineering, IEEE Transactions on 47 (9)
CE
(2000) 1185–1194.
[26] C. Berthomier, X. Drouot, M. Herman-Stoca, P. Berthomier, J. Prado,
AC
D. Bokar-Thire, O. Benoit, J. Mattout, M.-P. dOrtho, Automatic analysis of single-channel sleep eeg: Validation in healthy individuals, Sleep 30 (11) (2007) 15871595.
[27] The dreams subjects database. URL http://www.tcts.fpms.ac.be/ devuyst/Databases/DatabaseSubjects/ 38
ACCEPTED MANUSCRIPT
[28] I. Selesnick, Wavelet transform with tunable Q-factor, Signal Processing, IEEE Transactions on 59 (8) (2011) 3560–3575.
CR IP T
[29] S. Patidar, R. B. Pachori, U. R. Acharya, Automated diagnosis of coro-
nary artery disease using tunable-Q wavelet transform applied on heart rate signals, Knowledge-Based Systems 82 (0) (2015) 1 – 10.
[30] A. R. Hassan, M. I. H. Bhuiyan, Computer-aided sleep staging using
AN US
complete ensemble empirical mode decomposition with adaptive noise and bootstrap aggregating, Biomedical Signal Processing and Control 24 (2016) 1 – 10.
[31] K. Fukunaga, Introduction to Statistical Pattern Recognition, 2nd Edi-
M
tion, 2009.
[32] C.-S. Huang, C.-L. Lin, L.-W. Ko, S.-Y. Liu, T.-P. Su, C.-T. Lin,
ED
Knowledge-based identification of sleep stages based on two forehead electroencephalogram channels, Frontiers in Neuroscience 8 (263).
PT
[33] A. Hassan, S. Bashar, M. Bhuiyan, On the classification of sleep states by means of statistical and spectral features from single channel electroen-
CE
cephalogram, in: Advances in Computing, Communications and Informatics (ICACCI), 2015 International Conference on, 2015, pp. 2238–
AC
2243.
[34] A. R. Hassan, M. A. Haque, Computer-aided gastrointestinal hemorrhage detection in wireless capsule endoscopy videos, Computer Methods and Programs in Biomedicine 122 (3) (2015) 341 – 353.
39
ACCEPTED MANUSCRIPT
[35] L. Breiman, Bagging predictors, Machine Learning 24 (2) (1996) 123– 140.
Systems Magazine, IEEE 6 (3) (2006) 21–45.
CR IP T
[36] R. Polikar, Ensemble based systems in decision making, Circuits and
[37] A. R. Hassan, Computer-aided obstructive sleep apnea detection using normal inverse gaussian parameters and adaptive boosting, Biomedical
AN US
Signal Processing and Control 29 (2016) 22 – 30.
[38] L. G. Doroshenkov, V. A. Konyshev, S. V. Selishchev, Classification of human sleep stages based on eeg processing using hidden markov models, BioMedical Engineering 41 (1) (2007) 24–28.
M
[39] S.-F. Liang, C.-E. Kuo, Y.-H. Hu, Y.-S. Cheng, A rule-based automatic sleep staging method, Journal of Neuroscience Methods 205 (1) (2012)
ED
169 – 176.
[40] C. Vural, M. Yildiz, Determination of sleep stage separation ability of
PT
features extracted from eeg signals using principal component analysis, Journal of Medical Systems 34 (2010) 83–89.
CE
[41] Y.-L. Hsu, Y.-T. Yang, J.-S. Wang, C.-Y. Hsu, Automatic sleep stage recurrent neural classifier using energy features of eeg signals, Neuro-
AC
computing 104 (0) (2013) 105 – 114.
[42] A. R. Hassan, S. K. Bashar, M. I. H. Bhuiyan, Automatic classification of sleep stages from single-channel electroencephalogram, in: India Conference (INDICON), 2015 Annual IEEE, 2015, pp. 1–6. 40
ACCEPTED MANUSCRIPT
[43] A. Iranzo, J. L. Molinuevo, J. Santamara, M. Serradell, M. J. Mart, F. Valldeoriola, E. Tolosa, Rapid-eye-movement sleep behaviour disorder
The Lancet Neurology 5 (7) (2006) 572 – 577.
CR IP T
as an early marker for a neurodegenerative disorder: a descriptive study,
[44] A. R. Hassan, M. I. H. Bhuiyan, A decision support system for automatic
sleep staging from eeg signals using tunable Q-factor wavelet transform
– 118.
AN US
and spectral features, Journal of Neuroscience Methods 271 (2016) 107
[45] R. G. Baraniuk, D. L. Jones, A signal-dependent time-frequency representation: optimal kernel design, IEEE Transactions on Signal Processing 41 (4) (1993) 1589–1602.
M
[46] D. L. Jones, R. G. Bariniuk, An adaptive optimal-kernel time-frequency representation, in: 1993 IEEE International Conference on Acoustics,
ED
Speech, and Signal Processing, Vol. 4, 1993, pp. 109–112 vol.4.
PT
[47] B. Boashash, S. Ouelha, An improved design of high-resolution quadratic time-frequency distributions for the analysis of nonstationary multicomponent signals using directional compact kernels, IEEE
AC
CE
Transactions on Signal Processing 65 (10) (2017) 2701–2713.
41