A decision support system for automated identification of sleep stages from single-channel EEG signals

A decision support system for automated identification of sleep stages from single-channel EEG signals

Accepted Manuscript A decision support system for automated identification of sleep stages from single-channel EEG signals Ahnaf Rashik Hassan, Abdul...

698KB Sizes 0 Downloads 58 Views

Accepted Manuscript

A decision support system for automated identification of sleep stages from single-channel EEG signals Ahnaf Rashik Hassan, Abdulhamit Subasi PII: DOI: Reference:

S0950-7051(17)30206-X 10.1016/j.knosys.2017.05.005 KNOSYS 3902

To appear in:

Knowledge-Based Systems

Received date: Revised date: Accepted date:

23 December 2016 5 May 2017 6 May 2017

Please cite this article as: Ahnaf Rashik Hassan, Abdulhamit Subasi, A decision support system for automated identification of sleep stages from single-channel EEG signals, Knowledge-Based Systems (2017), doi: 10.1016/j.knosys.2017.05.005

This is a PDF file of an unedited manuscript that has been accepted for publication. As a service to our customers we are providing this early version of the manuscript. The manuscript will undergo copyediting, typesetting, and review of the resulting proof before it is published in its final form. Please note that during the production process errors may be discovered which could affect the content, and all legal disclaimers that apply to the journal pertain.

ACCEPTED MANUSCRIPT

Highlights

posed.

CR IP T

• A single channel EEG based automated sleep scoring method is pro-

• A novel signal processing technique, namely TQWT is employed for sleep staging. • We introduce bagging to classify sleep stages.

AN US

• Efficacy of the method is confirmed by statistical and graphical analyses.

• The performance of the proposed scheme, compared to the existing

AC

CE

PT

ED

M

ones is promising.

1

ACCEPTED MANUSCRIPT

A decision support system for automated identification of sleep stages from single-channel EEG signals

CR IP T

Ahnaf Rashik Hassana,b,∗, Abdulhamit Subasic a

AN US

Institute of Biomaterials and Biomedical Engineering, University of Toronto, Toronto, ON, Canada b Department of Electrical and Electronic Engineering, Bangladesh University of Engineering and Technology, Dhaka-1000, Bangladesh. c Computer Science Department, College of Engineering, Effat University, Jeddah 21478, Saudi Arabia.

Abstract

A decision support system for automated detection of sleep stages can alleviate the burden of medical professionals of manually annotating a large

M

bulk of data, expedite sleep disorder diagnosis, and benefit research. Moreover, the implementation of a sleep monitoring device that is low-power and

ED

portable requires a reliable and successful sleep stage detection scheme. This article presents a methodology for computer-aided scoring of sleep stages us-

PT

ing singe-channel EEG signals. EEG signal segments are first decomposed into sub-bands using tunable-Q wavelet transform (TQWT). Four statisti-

CE

cal moments are then extracted from the resulting TQWT sub-bands. The proposed scheme exploits bootstrap aggregating (Bagging) for classification.

AC

Efficacy of the feature generation scheme is evaluated using intuitive, statistical, and Fisher criteria analyses. Furthermore, the efficacy of Bagging is evaluated using out-of-bag error analysis. Optimal choices of Bagging ∗

Corresponding Author: Ahnaf Rashik Hassan, Phone: +1-647-3367757, Email: [email protected]

Preprint submitted to Knowledge-Based Systems

May 8, 2017

ACCEPTED MANUSCRIPT

and TQWT parameters are explicated. The proposed methodology for automated sleep scoring is tested on the benchmark Sleep-EDF database and

CR IP T

DREAMS subjects database. Our methodology achieves 92.43%, 93.69%, 94.36%, 96.55%, and 99.75% accuracy for 2-state to 6-state classification of sleep stages on Sleep-EDF database. Experimental results show that the

algorithmic performance of the automated sleep scoring technique presented herein achieves better performance as compared to the state-of-the-art sleep

AN US

staging algorithms. Besides, the proposed scheme performs equally well for two sleep scoring standards, namely- AASM and R&K. Moreover, the proposed decision support system yields high success rate for identifying sleep

states REM and non-REM 1. It can be anticipated that owing to its use of only one channel of EEG signal, the proposed method will be suitable for

M

device implementation, eliminate the onus of medical professionals of anno-

diagnosis.

ED

tating a large volume of recordings manually, and expedite sleep disorder

Keywords: Sleep, EEG, Classification, Wavelet, Statistical moments.

PT

1. Introduction

CE

Sleep scoring is the first step of sleep research and various sleep disorder diagnosis. Traditionally, sleep scoring is performed by expert sleep techni-

AC

cians based on visual observation of signals. Sleep in humans is composed of cyclic alternations of non-rapid eye movement (NREM) and rapid eye movement (REM) sleep states [1]. NREM phase is yet again classified into four states (S1-S4). Stages S4 and S3 are collectively termed as slow wave sleep owing to the dominance of slow oscillations which emanate from neocortex [2]. 3

ACCEPTED MANUSCRIPT

Again, sleep stages S2 and S1 together are termed as shallow sleep [3]. For visual sleep scoring, nocturnal polysomnography (PSG) is scored by expert

CR IP T

sleep scorers in accordance with Rechtschaffen and Kales’s (R&K) guidelines [4] or a recent set of recommendations by American Academy of Sleep Medicine (AASM) [5]. In R&K guidelines, the recordings are annotated in

any one of the 6 stages- S1-S4, Awake (AWA), and REM. The main alterations of AASM guidelines consist of derivations of EEG, “movement time”

AN US

state abolition, combination of stages S4 and S3 into N3, modification of

some context rules in order to make them simpler, and filter setting guidelines [6] [7].

Sleep state detection by examining the EEG signals visually is problematic for many reasons. Sleep stage assessment by technicians is dependent

M

on human resources, which is expensive. The process is also dependent on rater’s experience and expertise [8]. Moreover, the huge bulk of data that

ED

are inspected per patient make sleep scoring by human experts burdensome and subject to misclassification owing to fatigue [9]. In clinical scenarios,

PT

some cases may demand urgent diagnosis [10] [11]. Manual sleep scoring, due to its time-consuming process, cannot deal with such situations. This

CE

failure, in turn, leads to belated sleep disorder diagnosis. Again, given its consuming nature, human expert based sleep stage screening is unsuitable to handle large data-sets. As a result, large-scale population studies in sleep

AC

research are greatly hindered by this caveat [12] [13]. So, it is evident that there is a great need of an automatic sleep scoring system. Now one might ask- why must the sleep scoring scheme be single-channel based? Developing a wearable, unobtrusive, portable, non-invasive, yet low-power

4

ACCEPTED MANUSCRIPT

device is essential for in-home sleep monitoring. The device must be designed and implemented in such a manner that it will reliably conduct a preliminary

CR IP T

test of the user at home without the requiring a doctor. The development of such a wearable device has gained the interest of sleep research community. A reliable monitoring device of this kind will be instrumental in early diagno-

sis of sleep disordered breathing, such as- sleep apnea and other sleep related disorders. With a view to ensuring portability, convenience of the user, and

AN US

wearability, the automated sleep stage identification scheme employed by the device must be single-channel based. This will also make sure that the battery has an enhanced life and the size of the device is comfortable and convenient for the user. The development and subsequent implementation of such as sleep monitoring system, thus, solely relies on a single-channel based

M

reliable computerized sleep staging scheme.

Various prior studies have attempted to perform sleep stage classifica-

ED

tion [14]. Diykh et al. [15] utilized time domain statistical features and structural graph similarity for feature extraction and K-means in order to

PT

classify sleep stages. Kouchaki et al. [16] used tensor based singular spectrum analysis for computerized identification of sleep-EEG signals. Hassan

CE

et al. [17] put forward a computerized sleep staging scheme employing empirical mode decomposition (EMD) and adaptive boosting. Riaz et al. [18] employed spectral and temporal features extracted in the EMD domain to

AC

devise their computerized sleep scoring scheme. Owing to the use of EMD, both of the aforementioned schemes suffer from mode-mixing problem. Tsinalis et al. [19] implemented a time-frequency analysis based feature extraction and a stacked sparse autoencoders based classification scheme to design

5

AN US

CR IP T

ACCEPTED MANUSCRIPT

Figure 1: A scheme of our computer-aided sleep stage scoring framework.

their system for automatic scoring of sleep-EEG. Peker [20] utilized a combination of dual-tree complex wavelet transform (DT-CWT) and Taguchi

M

based complex-valued neural network for computerized sleep staging from one channel of EEG signals.

ED

In the present study, we present a system for automated scoring of sleep states from single-channel EEG signals. The framework proposed herein is

PT

demonstrated in Fig. 1. The EEG signal segments belonging to various sleep states are first decomposed into sub-bands using a data-driven signal

CE

analysis technique, namely- tunable Q-factor wavelet transform (TQWT). Four statistical moments, namely- mean, variance, skewness, and kurtosis are then extracted from each of the sub-bands. The test and the train ma-

AC

trices are formed afterwards. For feature selection and for determining the statistical significance of the differences of the selected features among all the sleep states, Kruskal-Wallis one-way analysis of variance (Kruskal-Wallis ANOVA) is performed. The features that do not pass Kruskal-Wallis ANOVA

6

ACCEPTED MANUSCRIPT

test are removed and the feature matrices are then fed into the classifier. The classifier used in this work is an eminent ensemble-meta learning based clas-

CR IP T

sification model, namely- bootstrap aggregating (Bagging). Influences and optimal choices of various bagging and TQWT parameters are inspected.

Our experimental findings evince that the algorithmic performance of our

scheme is promising. Algorithmic performance evaluation against the stateof-the-art studies suggest that the proposed sleep scoring scheme performs

AN US

better than or comparable to existing studies in terms of various evaluation metrics. Furthermore, the proposed scheme gives higher detection performance for REM and S1 stages than many of the existing computer-aided sleep scoring schemes.

The organization of the remainder of the article is described below. Sec-

M

tion 2 elucidates the proposed feature extraction framework, investigates its efficacy, describes the statistical analysis that has been performed, and de-

ED

scribes the classifier used in the proposed method. Section 3 expounds the experimentations that have been conducted, presents their results, and expli-

PT

cates their significance. Finally, Section 4 makes some concluding remarks, points to some of the future directions of this work, and brings this article

CE

to conclusion.

AC

2. Materials and methods In this study, Sleep-EDF database available at Physionet and DREAMS

Subjects database have been utilized to conduct the experiments.

7

ACCEPTED MANUSCRIPT

10

15

20

0

5

10

15

20

0

5

10

15

20

0

5

10

15

20

0

5

10

15

0

5

10

15 time (seconds)

25

30

25

30

25

30

25

30

20

25

30

20

25

30

M

50 0 −50

5

CR IP T

50 0 −50

100 0 −100

AWA

REM

50 0 −50

0

AN US

S1

100 0 −100

Amplitude (µV) S4 S3

S2

EEG Epochs of Various Sleep States 100 0 −100

ED

Figure 2: Examples of EEG signal segments of the six sleep states.

2.1. Experimental data

PT

2.1.1. Sleep-EDF data-set

Sleep-EDF is an open-source, benchmark, and extensively utilized database

CE

in sleep scoring literature [21], [22], [23], [24]. Eight Caucasian female and male subjects took part in the study. The subjects’ ages ranged from 21-

AC

35 years. At the time of the study, none of the subjects was not taking any medication. Sleep-EDF has eight recordings. The EEG data have been grouped into two subsets (termed as sc∗ and st∗ ). The first set of EEG data (sc4112e0, sc4102e0, sc4012e0, sc4002e0) have been recorded in a study in 1989. The study subjects were ambulatory healthy. The second set of 8

ACCEPTED MANUSCRIPT

EEG data (st7132j0, st7121j0, st7052j0, st7022j0) have been recorded in a study in 1994. The participants of this study felt slight difficulty in falling

CR IP T

asleep even though all the subjects where health. Data collection has been performed in 24 hours of the subjects’ daily lives. In this subset, nocturnal EEG data have been collected from the four subjects in a hospital with the

help of a miniature telemetry system [25]. For both of the subsets, data of two EEG channels- Fpz-Cz and Pz-Oz have been collected. The sampling

AN US

frequency of the EEG signals is 100 Hz. Prior studies [21], [22], [23], [26] have demonstrated that EEG data from Pz-Oz channel give better algorithmic performance than Fpz-Cz channel. So, data from Pz-Oz channel have been used herein. As per R&K criteria, EEG recordings of both of the subsets have been annotated by an experienced sleep technician on 30s basis.

M

Therefore, the duration of each epoch is 30s or 3000 samples. The epochs have been annotated by sleep technicians as- AWA, REM, S1, S2, S3, S4,

ED

Movement Time, or ‘Unscored’. Fig. 2 shows some EEG signal segments of all the sleep stages. We also consider five cases of classification for both of

PT

the databases used in this study as Table 1 suggests. The six-state case is the R&K scoring itself- AWA, S1-S4, REM. The 5-state case combines S4

CE

and S3 of 6-state as slow wave sleep. The 4-state case combine S2 and S1 of 5-state as Table 1 indicates. The 3-state and 2-state cases consist of- REM, AWA, NREM and Sleep (REM and S1-S4), AWA respectively. Since AASM

AC

guideline combines S3 and S4 into N3 to consider 5 stages of sleep, case I is not considered for AASM scoring. The epoch distribution of each of the states of sleep of this database is shown in Table 2. The total EEG signal segments from Sleep-EDF used in this study is 15,188.

9

ACCEPTED MANUSCRIPT

Table 1: Cases considered for classification in this study.

Number of Classes

Constituent Sleep States

I

6

AWA, REM, S1, S2, S3, S4

II

5

AWA, REM, S1, S2, SWS(S3-S4)

III

4

AWA, REM, S1-S2, SWS(S3-S4)

IV

3

AWA, REM, NREM (S1-S4)

V

2

AWA, Sleep (REM & NREM)

AN US

CR IP T

Case

Table 2: Description of epochs of various sleep states of sleep-EDF database.

S1

S2

S3

S4

REM

Total

8055

604 3621 672

627

1609

15188

M

Epochs

AWA

ED

2.1.2. DREAMS Subjects database

The second database that has been used in this work is DREAMS Subjects database. 20 healthy subjects (16 females, 4 males, 20-65 years old)

PT

who were not taking any medication, participated in this study and 20 nocturnal PSG have been collected from the study participants. The study has

CE

been conducted in the sleep lab of Andr Vsale hospital (Montigny-le-Tilleul, Belgium). Each recording is 7-9 hours in duration. DREAMS Subjects has

AC

168.81 hours of EEG data. The number of EEG segments or epochs extracted from this database is 20,257. A digital 32-channel polygraph (BrainnetTM System of MEDATEC, Brussels, Belgium) has been employed to collect the recordings. DREAMS Subjects database contains one submental EMG chan-

10

ACCEPTED MANUSCRIPT

nel, 3 channels of EEG data (C3-A1 or CZ-A1, O1-A1, and FP1-A1) and 2 channels of EOG data (P18-A1, P8-A1) have been recorded. The data have

CR IP T

been stored in European data format (EDF). The sampling rate of the signals was 200 Hz. Annotations in accordance with both R&K and AASM criteria have been done by an expert sleep technician of the sleep lab. R&K criteria considers sleep states- AWA, REM, S1-S4, ‘sleep stage movement’

and ‘unknow sleep stage’. The unknown sleep stages and movement times

AN US

have been excluded from this study. Again, as per AASM standards, epochs have been annotated in five states- AWA, N1, N2, N3, REM [27]. To carry out the experiments in this study, 50 percent of the Sleep-EDF database has been chosen at random as training data and the other half of the database as testing data. Epochs of each sleep stage were divided

M

equally to ensure that both the training and the testing sets contain epochs of all the sleep stages. The same strategy is adopted for DREAMS Subjects

ED

database as well. In this manner, we have made sure that for each of the two databases, the entire database is used either for training or testing, yet

PT

never both simultaneously. This data distribution decreases the chance of overfitting as well. The process to divide the databases is repeated 20 times.

CE

The mean of the performance metrics of this 20 runs have been presented in this paper.

AC

2.2. Tunable-Q factor wavelet transform Tunable-Q factor wavelet transform (TQWT) is a newly developed signal

decompositions technique that is data-adaptive and flexible. That is why, it has been proven to be an ideal technique to process oscillatory signals [28]. TQWT ensures flexibility by varying its parameters- Q-factor (Q), no. of 11

ACCEPTED MANUSCRIPT

H0j (ω)

LP Scaling αj

(a)

s[n]

H1j (ω)

LP Scaling αj−1

HP Scaling β

dj [n]

AN US

(b)

cj [n]

CR IP T

s[n]

Figure 3: A scheme depicting how an input signal s[n] is decomposed by TQWT up to jth level. (a) the low-pass sub-band signal cj [n] and (b) the high-pass sub-band signal dj [n].

decomposition levels (J), and over-sampling rate (r). Q controls the wavelet’s

ED

M

oscillation. The mathematical formulation of Q is given below.

CE

PT

where

Q=

fc (j) BW (j)

fc (j) = αj

2−β 4α

1 BW (j) = βαj−1 π 2

(1)

(2)

(3)

and j = 1, 2, ..., J. For decomposition of J levels, TQWT converts an

AC

input signal s[n] into J + 1 sub-bands. For this conversion, a filer bank with two channels is used iteratively. TQWT applies the filter bank to the lowpass sub-band. Input s[n] is decomposed into c0 [n] and d1 [n] in each of the stages. Here, c0 [n] denotes the low-pass sub-band with sampling frequency

12

ACCEPTED MANUSCRIPT

of αfs . d1 [n] indicates the high-pass sub-bands with sampling frequency of βfs . The scaling factors are denoted by α and β and fs is the sampling rate

CR IP T

of s[n]. Low-pass filter H0 (ω) with low-pass scaling α is applied to generate c0 [n]. d1 [n] is created by H1 (ω) and high-pass scaling β. To make sure that

TQWT has perfect reconstruction and no redundancy, the values of α and

β must be: 0 < α < 1, 0 < β ≤ 1 and α + β > 1. For the implementation of TQWT, we have used the toolbox provided by Selesnick [28].

AN US

Fig. 3 presents a schematic outline of how a signal s[n] is processed by TQWT up to jth level to produce cj [n] and dj [n]. H0j (ω) is the frequency response of the low-pass sub-bad and H1j (ω) is the frequency response of the high-pass sub-band created after j-levels. Their mathematical formulations are given below.

PT

ED

 0,

M

(J)

H0 (ω) =

 Q   J−1 H0 (ω/αm ), if |ω| ≤ αJ π m=0

(J)

CE

H1 (ω) =

AC

where

if α π ≤ |ω| ≤ π

 Q  m  H1 (ω/αJ−1 ) J−2  m=0 H0 (ω/α ),      if(1 − β)αJ−1 π ≤ |ω| ≤ αJ−1 π    0,      for otherω ∈ [−π, π]

ω + (β − 1)π ) α+β−1 απ − ω H1 (ω) = θ( ) α+β−1

H0 (ω) = θ(

(4)

J

(5)

(6) (7)

θ(ω) denotes Daubechies filter’s frequency response whose mathematical 13

ACCEPTED MANUSCRIPT

formulation is given below. (8)

CR IP T

p θ(ω) = 0.5(1 + cos(ω)) 2 − cos(ω), |ω| < π TQWT parameters r and Q are related to α and β as: r=

β , 1−α

Q=

2−β β

2.3. Advantages of TQWT

(9)

AN US

The use of TQWT in the proposed scheme is motivated by various advantages of TQWT over conventional signal processing techniques such as empirical mode decomposition (EMD) and Fourier transform. • The Q-factor of the wavelet must be low as the wavelet transform at-

M

tempts to process less oscillatory signals. On the contrary, the Q-factor of the wavelet must be high as the wavelet transform attempts to pro-

ED

cess highly oscillatory (e.g. EEG) signals. Nevertheless, except for the continuous wavelet transform, most wavelet transforms cannot tune the Q-factor [28]. TQWT eradicates this limitation by tuning its param-

PT

eter Q. Therefore, it has been successful in analysis and classification

CE

of EEG.

• Feature generation schemes that use Fourier transform based features,

AC

assume that the signal is stationary and linear even though EEG is highly non-stationary and nonlinear. In spite of their data-adaptive nature, EMD and almost all of its variations are affected by modemixing. TQWT, on the contrary, does not have this caveat. Therefore, it is more suitable for EEG data analysis. 14

ACCEPTED MANUSCRIPT

• TQWT has been proven to be instrumental in analyzing various biosignals successfully of late [29].

CR IP T

• The transfer functions of TQWT filters are rational, contributing to its less computational burden and efficiency. This also makes the frequency domain specification easier [28].

• The advantage of TQWT over commonly used time-frequency based

AN US

transforms is that TQWT is highly flexible. In other words, TQWT can alter its Q-factor to characterize the morphology of the signal in

the frequency domain in a better manner. Other time-frequency based transforms can not offer this flexibility.

• Last but not the least, wavelet transform’s perfect reconstruction prop-

M

erty is inherited by TQWT.

ED

The advantages mentioned above establish TQWT as a prominent signal decomposition scheme of analyzing oscillatory EEG signals and motivate its use in our framework. Histograms of TQWT sub-band of the six sleep phases

PT

in Fig. 4 further corroborate with the appealing characteristics of TQWT. It is to be noted in Fig. 4 that the shapes of histograms of the EEG signals are

CE

different from each other. For instance, histograms corresponding to S1 and S2 are somewhat skewed to the right as opposed to the histograms of other

AC

sleep stages. Also, the dispersion of the histograms are different from each other. Another key difference among the histograms in Fig. 4 is that the ranges of amplitudes of EEG signals in different sleep stages are different as well. The histograms shapes are different for the six phases of sleep in Fig. 4, conforming yet again the efficacy of TQWT for analyzing sleep-EEG. 15

ACCEPTED MANUSCRIPT

S2 State

40

40

40

30 20

Occurrence

50

30 20 10

0 −40

−20

0 Voltage (µV) S4 State

20

−50

0 Voltage (µV) AWA State

50

0 −200

100

60

50

50

40

40

20 10

Occurrence

60

40 Occurrence

50

30

30 20 10

−100

0 Voltage (µV)

100

200

0 −200

−100

0 Voltage (µV) REM State

100

200

30 20 10

−100

0 100 Voltage (µV)

200

300

0 −100

−50

AN US

0 −200

20 10

0 −100

40

30

CR IP T

10

Occurrence

S3 State

50

Occurrence

Occurrence

S1 State 50

0 Voltage (µV)

50

100

Figure 4: TQWT sub-bands’ histograms of the sleep states. Significant differences of histogram shapes are observed among the six stages of sleep. Q = 2, r = 3, J = 15 in this example.

M

2.4. Statistical moment based features

Once we have obtained the TQWT sub-bands, the next step is to extract

ED

four statistical moments to be used as features from each of the sub-bands as Fig. 1 evinces. The use of statistical moments as features in this study

PT

is motivated by their successful and widespread use in physiological signal analysis [30]. As these measures signify the asymmetry, peakedness, and dis-

CE

persion of a data distribution, one can anticipate that statistical features will capture the underlying differences of the TQWT sub-bands corresponding to

AC

the six classes illustrated in Fig. 4. This hypothesis is validated by both the good discrimination power and classification performance of the features as we shall see in subsequent sections of this article. For a TQWT sub-band  with length ζ, T = τ1 , τ2 , τ3 , .., τN mean (θ), variance (σ 2 ), skewness (ι), and kurtosis (ν) can be formulated as shown in Table 3. 16

ACCEPTED MANUSCRIPT

Table 3: Description of the four features used in the proposed scheme.

Mathematical Formulation

Mean (θ)

θ= σ2 =

Skewness (ι)

ι=

1 ζ

ν=

1 ζ

i=1 τi

2 τi − θ  3

i=1

PN

i=1

PN

i=1

τi −θ σ



τi −θ σ

4

AN US

Kurtosis (ν)

PN

PN

1 ζ

Variance (σ 2 )

1 ζ

CR IP T

Feature Name

2.5. Separability evaluation

The efficacy of statistical moments extracted from TQWT sub-bands is yet again demonstrated herein by separability analysis. Our feature genera-

M

tion scheme’s effectiveness is now evaluated using Fisher criteria [31]. Unlike most of the separability metrics, Gaussian assumption of the data in consid-

ED

eration is not required for Fisher criteria [31]. The two Fisher criteria are

PT

expressed mathematically as follows. J1 =

tr(Sm ) tr(Sw )

J2 = tr(Sw−1 Sm )

(10) (11)

CE

where Sw and Sm evince the within-class and the between-class scatter

matrices respectively and tr(S) denotes the trace of S. The larger the values

AC

of J2 and J1 are, the features in consideration are more separable in feature space. Table 4 compares J1 and J2 values of our method with two approaches presented in [32]. It is evident from Table 4 that the J1 and J2 values of

our scheme are higher. This indicates that statistical moments from TQWT sub-bands give high separability among the six sleep phases. 17

ACCEPTED MANUSCRIPT

Table 4: Comparison of J1 and J2 of various schemes on Sleep-EDF.

J1

Power spectrum density features [32] STFT and RVM [32] Statistical features and CEEMDAN [30]

0.69

1.82

0.74

4.57

0.88

56.757

0.68

66.4

AN US

Spectral features from sleep-EEG signals [33] Proposed Method

2.6. Feature selection

J2

CR IP T

Methods

1.3147 125.5887

M

With a view to ascertaining if the values of the four statistical features differ significantly in the six sleep stages (AWA, S1-S4, and REM) and if

ED

the discrimination ability of the four statistical features extracted from the TQWT domain are significant statistically, a Kruskal-Wallis one-way analy-

PT

sis of variance is carried out. An advantage of Kruskal-Wallis test over other hypothesis testing schemes is that it does not require the underlying data

CE

distribution to be normal [34]. Kruskal-Wallis ANOVA is performed at a confidence level of 95%. Consequently, a difference is regarded to have sig-

AC

nificance if p < α(= 0.05) is found. The result of Kruskal-Wallis ANOVA is presented in Table 5. We can see that features computed from some of the sub-bands do not pass the hypothesis test. Before classification using the test and the train matrices, these features are removed. Additionally, we can see from Table 5 that of the features that pass the test, almost all of them

18

ACCEPTED MANUSCRIPT

Table 5: p-Values of θ, σ 2 , ι, and ν across various TQWT sub-bands for case I. Statistically insignificant features are shown in bold.

PT

ι 0.002586 0.003781 0.000514 0.0157 0.1539 0.002926 0.009327 0.002203 0.00999 0.003088 0.005488 0.1265 0.6956 0.008618 0.001954 0.002598 0.007907 0.008866 0.003938 0.008379 0.0085 0.6936 0.006309 0.009433

ν 1.3926e-277 8.7060e-291 8.3727e-243 7.3642e-158 5.2075e-57 2.3317e-277 3.7052e-173 1.6636e-93 4.9284e-94 1.3895e-124 4.5837e-144 2.1401e-134 2.4296e-110 1.4980e-92 6.2755e-69 1.4292e-64 1.4344e-34 7.4337e-40 2.9448e-46 3.0988e-30 6.3235e-29 6.5229e-36 2.1707e-17 1.9109e-07

CR IP T

σ2 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0

M

AN US

θ 0.6762 0.002916 0.000821 0.625 0.000585 0.008186 0.00824 0.005205 0.000957 0.0009891 0.008398 0.004177 0.006163 0.002103 0.00026 0.000953 0.001756 0.007923 0.004003 0.00344 0.9744 0.00396 0.00723 0.000564

ED

Sub-band 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24

CE

give lower p-values evincing good inter-class variation of the features. Thus, the results of Kruskal-Wallis ANOVA confirm the inferences drawn from Fig.

AC

4 and from Fisher criteria evaluation that the four statistical moments can characterize the sleep stages successfully and the proposed feature extraction scheme is efficacious.

19

ACCEPTED MANUSCRIPT

Algorithm 1: Bagging Input: N number of training examples S = [(xi , yi )] of γ classes having labels yi ∈ ω, ω = {ω1 , ω2 , ...., ωγ } and i = 1, 2, 3, ....., N . Base

CR IP T

learner or weak classifier or learner WeakLearn. A fraction φ to form

bootstrapped training data-set. T , an integer that indicates the total iterations.

for t = 1, 2, ...., T do 1. Draw a bootstrapped replica St by taking φ percent of S at

AN US

random.

2. Call WeakLearn with St and get the hypothesis (classifier) νt . νt to the ensemble η. end

Test: For a test instance, x with unknown label:

M

1. Evaluate η = {ν1 , ν2 , ..., νT } on x.   1,

ED

2. vt,j =

 0,

if νt picks class ωj Otherwise

CE

PT

3. Find the total vote obtained by each class: Vj =

T X

vt,j

t=1

where j = 1, 2, .., γ

AC

4. Select the class that has obtained the highest number of votes as bagging’s output.

20

ACCEPTED MANUSCRIPT

2.7. Bootstrap aggregating Bootstrap aggregating (Bagging) is an ensemble-meta algorithm first pro-

CR IP T

posed in [35] which creates and combines multiple classification models to solve a particular classification problem. It is one of the simplest and most

intuitive to implement ensemble learning based classifier, even though it yields strikingly good performance [36]. Bagging is depicted Algorithm 1.

Bagging achieves diversity by employing original training data-set’s boot-

AN US

strapped replicas. First, subsets are taken at random with replacement from the original training data-set. Afterwards, each subset St drawn as such is utilized to train a different weak classifier, generally of the same type. In this

way, a hypothesis νt is obtained each time. This weak classifier or predictor is represented as WeakLearn in Algorithm 1. In this article, we use deci-

M

sion trees as our weak classifier or learning algorithm. Individually obtained hypotheses are combined afterwards by means of majority voting. For any

ED

given unlabeled test instance, the ensemble’s decision is the class chosen by most weak classifiers. The number of iterations T and WeakLearn parame-

PT

ters (e.g. number of leaves for decision tree) affect Bagging’s performance as Algorithm 1 evinces. It is to be noted that T is equal to the total number of

CE

decision trees (WeakLearn) since one tree is trained in each iteration. The optimal choice of T is important. If T is too small or too large, it will result

AC

in poor performance or computational inexpensiveness of the algorithm respectively [33]. We shall expound how T and number of leaves are selected

in the next section.

21

ACCEPTED MANUSCRIPT

OOB classification error vs. number of grown trees for case I to V 0.4 Case I Case II Case III Case IV Case V

CR IP T

0.3 0.25 0.2 0.15

AN US

Out−of−bag classification error

0.35

0.1 0.05

0

50

100 150 200 Number of decision trees

250

300

M

0

ED

Figure 5: Variation of OOB error with the number of decision trees.

3. Results and discussions

PT

Here we explicate the details and the outcomes of our experiments and expound some of the significance of the results.

CE

With a view to finding the optimal decision trees required to obtain the

best algorithmic performance and to examining how the accuracy value varies

AC

with the no. of trees, we find out the no. of decision trees which gives minimum out-of-bag (OOB) error. This process is repeated every time for cases I to V. The resulting values are used to classify the unlabeled test instances in each case. Fig. 5 depicts the variation of OOB error as we vary no. of trees from 1 to 300 for cases I to V. Fig. 5 manifests that after about 22

ACCEPTED MANUSCRIPT

LOOCV error variation with number of decision trees 18 Case I Case II Case III Case IV Case V

CR IP T

16

12 10 8

AN US

LOOCV error (%)

14

6 4

0

50

100 150 200 Number of decision trees

250

300

M

2

Figure 6: Variation of LOOCV error with number of decision trees reveals the prediction

ED

power of the classifier in all the cases.

PT

20 decision trees, the OOB error almost becomes constant. However, to make sure that the algorithm yields the best performance, no. of trees for which

CE

OOB error is minimum is noted every time for cases I to V. Afterwards, when classifying the test examples of each case, these values are used. For cases I to V, the optimal no. of decision trees are- 285, 265, 153, 295, and 188

AC

respectively.

Fig. 5 also manifests the prediction power of bagging for the proposed

feature extraction scheme. In fact, prior studies have empirically demonstrated that OOB estimation is tantamount to using a test data-set having

23

ACCEPTED MANUSCRIPT

Variation of accuracy values (%) with the variation of r for cases I−V. 98

CR IP T

94 92 90 88

Case V Case IV Case III Case II Case I

86 84

4

6

8

10

AN US

Accuracy in percentage

96

12

14

16

18

20

M

Value of r

Figure 7: Accuracy variation with r.

ED

equal size of a train data-set [35]. As OOB error estimates the true ensemble error unbiasedly, cross-validation or a separate test data-set to assess the

PT

prediction ability of the classification model is not needed anymore. In classification cases where insufficient data are available, this approach for analyzing

CE

the predictive performance is really helpful. The low OOB error of Fig. 5 also manifests good predictive capability of bagging. Furthermore, after 20

AC

trees, OOB error variation becomes trivial, indicating the robustness of our method. To further validate the prediction power of the classification model, we have conducted leave-one-out cross-validation (LOOCV) analysis. Fig. 6 shows that the proposed decision support system yields low LOOCV error for a wide range of number of decision trees for all the five cases of interest. 24

ACCEPTED MANUSCRIPT

Variation of accuracy values (%) with the variation of Q for cases I−V. 98 96

CR IP T Case V Case IV Case III Case II Case I

92 90 88 86 84 82 80 78

5

10

15

20

AN US

Accuracy in percentage

94

25

30

35

40

45

50

M

Value of Q−factor, Q

Figure 8: Accuracy variation with Q-factor Q.

ED

Thus, LOOCV further validates the prediction capability of the algorithm. The foregoing findings are later validated and confirmed by bagging’s high

PT

detection accuracy of classifying test data-set. Prior works on TQWT indicate that r ≥ 3 to circumvent excessive un-

CE

desirable wavelet ringing [28] and higher detection accuracy is obtained if r is lower [37]. The experimental findings of this study confirm this as well.

AC

Fig. 7 shows how the variation of the value of TQWT parameter r influences the algorithmic performance of the proposed sleep scoring framework. It is clear from Fig. 7 that the highest detection accuracy is obtained when the values of r are set to- 4, 5, 4, 4, and 5 for cases I to V. So, these values of r are used in this study. Again, for Q, the experiments indicate that the 25

ACCEPTED MANUSCRIPT

Variation of accuracy values (%) with the number of decomposition levels, J for cases I−V. 98

CR IP T

94 92 90 88

Case V Case IV Case III Case II Case I

AN US

Accuracy in percentage

96

86 84 82

5

10

15

20

25

30

M

Value of number of decomposition levels, J

Figure 9: Accuracy variation with number of decomposition levels J.

ED

highest detection accuracy is obtained when Q is set to 2, 2, 2, 3, and 1 respectively for cases I to V. This is confirmed by many experimental obser-

PT

vations wherein the values of the other two parameters J and r have been varied. Fig. 8 serves as a pictorial demonstration of this fact. The detection

CE

accuracy vs. J does not demonstrate a decreasing trend like Q and r as Fig. 9 evinces. Here the value of J is varied in the range of 1 to 30 in step of one

AC

setting the optimal values of r and Q in each of the five cases to find out the values of J for which detection accuracy is maximum. The highest detection accuracy is obtained when the value of J is set to 23, 21, 22, 14, and 15.

The parameter value choices have been confirmed and validated from many experimentations. Thus, it can be concluded that the aforementioned values 26

ACCEPTED MANUSCRIPT

Table 6: Confusion matrix for case I.

Predicted Class S3

S1

113

43

0

S2

3

1682

23

S3

0

141

116

S4

0

5

27

AWA

27

4

0

REM

20

84

S4

AWA

REM

Sen

CR IP T

S2

0

56

90

37.42%

8

58

37

92.93%

53

26

0

34.52%

264

18

0

84.35%

0

3971

26

98.61%

1

38

662

82.34%

AN US

Actual Class

S1

0

Table 7: Confusion matrix for case II.

Predicted Class

S3-S4

AWA

REM

Sen

43

0

53

89

38.74%

S2

5

1641

54

70

41

90.66%

S3-S4

0

104

513

33

0

79.04%

AWA

28

6

0

3971

23

98.61%

REM

24

84

1

41

655

81.47%

CE

PT

Actual Class

117

ED

S1

S2

M

S1

are the optimal values of TQWT parameters.

AC

The confusion matrices presented in Tables 6-10 give further insights on

the algorithmic performance of our scheme for cases I to V. It is clear from the matrices that in general, as classes increase, the detection accuracy of each sleep stage decreases. This is also reflected on the overall accuracy

27

ACCEPTED MANUSCRIPT

Table 8: Confusion matrix for case III.

Predicted Class AWA

REM

Sen

S1-S2

1879

60

85

89

88.97%

S3-S4

101

515

34

0

79.35%

AWA

63

0

3945

20

97.96%

REM

157

1

32

615

76.49%

CR IP T

S3-S4

AN US

Actual Class

S1-S2

Table 9: Confusion matrix for case IV.

Predicted Class

Actual Class

AWA

REM

Sen

2592

88

83

93.88%

65

3940

23

97.84%

163

41

601

74.75%

ED

REM

AWA

M

S1-S4

S1-S4

CE

PT

Table 10: Confusion matrix for case V.

S1-S4, REM

AWA

Sen

S1-S4, REM

3467

101

97.25%

AWA

113

3915

97.22%

AC

Actual Class

Actual Class

values for cases I to V. The proposed sleep scoring framework consistently yields high detection accuracy for all the sleep stages except for sleep stage S3 in case I as Table 6 indicates. One of the main reasons of poor per-

28

ACCEPTED MANUSCRIPT

Table 11: S1 detection accuracy of various methods for case I on Sleep-EDF data-set. The highest accuracy value is highlighted in bold.

S1

Doroshenkov et al. [38]

4.84%

Zhu et al. [21]

15.8%

Liang et al. [22]

18.75%

Liang et al. [39]

30%

AN US

CR IP T

Authors

33.70%

Ronzhina et al. [23]

36.17%

Hsu et al. [41]

36.70%

Proposed Method

37.42%

M

Vural et al. [40]

Table 12: S1 detection accuracy of various methods for case II on Sleep-EDF data-set.

AC

CE

PT

ED

The highest accuracy value is highlighted in bold.

Authors

S1

Zhu et al. [21]

15.8%

Liang et al. [22]

18.75%

Liang et al. [39]

30%

Hsu et al. [41]

36.70%

Proposed Method 38.74%

formance of computerized sleep staging algorithms is the resemblance of S1 state’s EEG signals with those of REM’s. In fact, S1 and REM are almost

29

ACCEPTED MANUSCRIPT

Table 13: Comparison of performance of various works on Sleep-EDF. Only the highest accuracy value (%) of each method is reported.

Fuzzy logic based iterative method [26]

CR IP T

Case ICase IICase IIICase IVCase V -

71.2

74.5

88.3

95.4

-

87.20

-

-

-

-

83.60

-

-

-

-

77.98

-

-

-

76.70

-

81.42

88.97

96.90

Temporal features and hidden Markov model [38] 61.08

-

-

-

-

-

-

-

-

Moment based features and adaptive boosting [42] 80.34 82.03

82.83

-

-

M

Methods

Frequency bands’ energy, neural network [41] Multiscale entropy, LDA [22]

AN US

Temporal features and hierarchical decision tree [39] PSD of single-channel EEG and ANN [23]

Hybrid features, Karhunen-Love transform [40]

69.98

85.57 86.53

87.49

89.77

95.05

Graph theory based features, SVM [21]

87.50 88.90

89.30

92.60

97.90

ED

Statistical features and bagging [33]

92.43 93.69 94.36 96.55 99.75

PT

Proposed Method

indistinguishable by visual inspection [21]. This is why, most S1 epochs are

CE

misclassified as REM by existing automatic sleep scoring methods. This is also evident from case I confusion matrix in Table 6 and case II confusion

AC

matrix in Table 7. An advantage of the algorithm propounded herein is that it classifies S1 state better than many of the prior studies. As we can see from the case I confusion matrix in Table 6, our method correctly classifies 37.42% S1 EEG epochs. The same scenario is seen in Table 7 as our method correctly classifies 38.74% S1 EEG epochs in case II. Moreover, in compar30

ACCEPTED MANUSCRIPT

ison with the extant state-of-the-art algorithms’ S1 detection performance, the proposed method’s S1 stage identification accuracy is significantly better

CR IP T

as Tables 11 and 12 suggest. Again, for sleep stage classification, REM phase is very important. REM stage consists of about 5-20% of an adult human’s

nocturnal sleep. Furthermore, detection of REM is a precondition for diagnosing many sleep ailments including narcolepsy, RBD etc. [43]. During

REM, muscle activity is observed to diagnose REM. It is a marker of many

AN US

neurological disorders such as Parkinsons disease. So, REM detection bears

particular significance in the context of sleep data analysis. It can be seen from Tables 6-9 that the proposed algorithm gives 82.34%, 81.47%, 76.49%, and 74.75% accuracy for REM detection for cases I to IV. In comparison, the methods put forward in [21] and [40] give 76.20% and 79.70% REM de-

M

tection accuracy respectively for case I as opposed to our method’s 82.34% REM detection accuracy in case I. So, REM phase detection success of our

ED

algorithm is comparable to or better than the existing studies. To get further insight on the efficacy of the proposed automated sleep

PT

scoring scheme, the performance of our method is compared with those of the existing works. For fair comparison, classification results obtained from

CE

Sleep-EDF are considered. Table 13 lists the accuracy values of various studies that use Sleep-EDF. All the accuracy values reported in this section and in Table 13 are the best detection accuracy for a particular method. It

AC

is clear from Table 13 that the method proposed in this work outperforms others in all the cases. To evaluate the effectiveness of the scheme propounded herein even fur-

ther, we compute Cohen’s kappa coefficient for each of the five cases consid-

31

ACCEPTED MANUSCRIPT

Table 14: Kappa coefficients of the proposed method against various published methods on Sleep-EDF data-set.

-

0.7968

0.81

II

0.7452

0.8

0.83

III

-

0.8125

0.83

IV

-

0.84

0.87

V

-

0.9

0.96

AN US

I

CR IP T

Case Liang et al. [22] Hassan et al. [30] Zhu et al. [21] Proposed Method 0.836

0.8543 0.863

0.8933 0.9435

ered in this work. Cohens kappa measures the agreement among annotators for categorial items statistically. It has been considered as a more robust measurement of performance than percent agreement since kappa considers the

M

coincidentally occurring agreements. Kappa coefficient ranging from 0.81-1 is regarded as almost perfect agreement, 0.61-0.80 as substantial agreement,

ED

0.41-0.60 as moderate agreement, 0.21-0.40 as fair agreement, 0-0.20 as slight agreement, and less than zero evinces no agreement. Table 14 evinces that

PT

the proposed scheme gives almost perfect agreement in all the cases of interest. Also, it gives higher accuracy values in comparison with prior works in

CE

all the cases except in case V. We now present the detection performance of our scheme on DREAMS

AC

Subjects. Table 15 lists the kappa coefficient and detection accuracy of our algorithm for AASM and R&K standards of sleep scoring. Even though the detection performance of our scheme is promising for all the cases, for case II in both of the scoring standards, accuracy values of case II are on the lower side. However, it can be observed that the proposed algorithm yields consis32

ACCEPTED MANUSCRIPT

Table 15: Algorithmic performance of the proposed method on DREAMS Subjects dataset.

R&K

76.39

74.39

Accuracy (%)

AASM

-

78.95

Cohen’s Kappa

R&K

0.785

0.78

Cohen’s Kappa

AASM

-

0.82

81.87

88.98

99.02

83.78

89.04

98.75

0.83

0.882

0.95

0.865

0.91

0.985

AN US

Accuracy (%)

CR IP T

Evaluation Metric Scoring Scheme Case I Case II Case III Case IV Case V

tently high detection accuracy in general in each of the five cases. Moreover, Table 15 reveals that automated sleep scoring gets more and more difficult for the algorithm with the increase of classes. This result supports the findings of

M

prior studies in Table 13. Furthermore, our scheme’s detection performance is quite similar in AASM and R&K scoring as Table 15 manifests. So, our

ED

sleep scoring system can be employed equally for AASM and R&K scoring scheme without any algorithmic changes or tuning of parameters, which is a big advantage of the system.

PT

Even though the suitability of TQWT for sleep data analysis has been demonstrated in [44], the proposed method, due to the use of the simple

CE

and computationally inexpensive statistical features, is particularly suitable for device implementation. apart from application in the clinical scenarios

AC

for assisting clinicians, one of the goals of developing an automated decision support system for sleep scoring is to implement the algorithm in a wearable and portable sleep monitoring device for in-home care. This necessitates the sleep scoring algorithm to be computationally as inexpensive as possible. The method proposed in this work not only gives high accuracy values, but it is 33

ACCEPTED MANUSCRIPT

computationally less expensive. This is owing to the fact that the proposed scheme involves computation of four simple statistical features, whereas the

CR IP T

Fourier domain based feature extraction scheme increases the computational complexity and expensiveness of the methodology presented in [44]. Thus,

the method presented in this work is more suitable for device implementation than the one presented in [44].

AN US

4. Conclusions and future directions

In this work, a methodology for computer-aided sleep scoring is presented. Sleep-EEG signal segments are first decomposed using TQWT and four statistical moments are extracted and employed as features. Bootstrap aggregating is employed for classification. Our experimental findings manifest that

M

statistical moment based features can capture the differences of various sleep states successfully. The influences of various Bagging and TQWT param-

ED

eters have been demonstrated. The algorithmic performance of our sleep scoring system as opposed to the existing ones is promising. Further, the

PT

proposed scheme yields good REM and S1 detection accuracy. Moreover, the proposed scheme equally performs well for R&K and AASM sleep scor-

CE

ing standards. In the future, we shall focus on hardware implementation of the proposed algorithm. A combination of various ensemble learning based

AC

classification models for the proposed feature extraction framework can be an interesting avenue of further research. Deep learning can also be employed to further meliorate classification performance in all the cases of interest. Future works can also employ time-frequency based methods [45] [46] such as multi-directional kernel methods [47] [9] and wrapper based feature selection 34

ACCEPTED MANUSCRIPT

techniques for computerized sleep scoring. To conclude, it can be said that the proposed scheme is efficient and efficacious and will alleviate the burden

References

CR IP T

of sleep technicians and expedite sleep disorder research and diagnosis.

[1] J. A. Hobson, Sleep is of the brain, by the brain and for the brain, Nature 437 (7063) (2005) 1254–1256.

AN US

[2] L. Marshall, H. Helgadottir, M. Molle, J. Born, Boosting slow oscillations during sleep potentiates memory, Nature 444 (7119) (2006) 610– 613.

[3] H. tieng Wu, R. Talmon, Y.-L. Lo, Assess sleep stage by modern signal

M

processing techniques, Biomedical Engineering, IEEE Transactions on

ED

62 (4) (2015) 1159–1168.

[4] A. Rechtschaffen, A. Kales, Manual of standardized terminology, techniques and scoring systems for sleep stages of human subjects, U. G. P.

PT

Office, Washington DC Public Health Service.

CE

[5] A. L. C. C. Iber, S. Ancoli-Israel, S. F. Quan, Westchester, usa: American academy of sleep medicine, The AASM Manual for the Scoring of

AC

Sleep and Associated Events: Rules, Terminology and Technical Specification.

[6] K. Samiee, P. Kovcs, M. Gabbouj, Epileptic seizure detection in longterm EEG records using sparse rational decomposition and local gabor binary patterns feature extraction, Knowledge-Based Systems (2016) –. 35

ACCEPTED MANUSCRIPT

[7] U. R. Acharya, H. Fujita, V. K. Sudarshan, S. Bhat, J. E. Koh, Application of entropies for automated diagnosis of epilepsy using EEG signals:

CR IP T

A review, Knowledge-Based Systems 88 (2015) 85 – 96. [8] S. B. Akben, A. Alkan, Visual interpretation of biomedical time series using parzen window-based density-amplitude domain transformation, PLOS ONE 11 (9) (2016) 1–13.

AN US

[9] B. Boashash, S. Ouelha, Automatic signal abnormality detection using

time-frequency features and machine learning: A newborn EEG seizure case study, Knowledge-Based Systems 106 (2016) 38 – 50. [10] S. Kara, F. Dirgenali, A system to diagnose atherosclerosis via wavelet transforms, principal component analysis and artificial neural networks,

M

Expert Systems with Applications 32 (2) (2007) 632 – 640.

ED

[11] B. Karlik, S. Aydin, An improved approach to the solution of inverse kinematics problems for robot manipulators, Engineering Applications

PT

of Artificial Intelligence 13 (2) (2000) 159 – 164. [12] H. G¨ ur¨ uler, A novel diagnosis system for parkinson’s disease using

CE

complex-valued artificial neural network with k-means clustering feature weighting method, Neural Computing and Applications (2016) 1–10.

AC

[13] H. M. Ertunc, H. Ocak, C. Aliustaoglu, Ann- and anfis-based multistaged decision algorithm for the detection and diagnosis of bearing faults, Neural Computing and Applications 22 (1) (2013) 435–446.

[14] A. Subasi, Automatic recognition of alertness level from EEG by using 36

ACCEPTED MANUSCRIPT

neural network and wavelet coefficients, Expert Systems with Applications 28 (4) (2005) 701 – 711.

CR IP T

[15] M. Diykh, Y. Li, P. Wen, Eeg sleep stages classification based on time domain features and structural graph similarity, IEEE Transactions on

Neural Systems and Rehabilitation Engineering PP (99) (2016) 1–1. doi:10.1109/TNSRE.2016.2552539.

AN US

[16] S. Kouchaki, S. Sanei, E. L. Arbon, D. J. Dijk, Tensor based singular

spectrum analysis for automatic scoring of sleep eeg, IEEE Transactions on Neural Systems and Rehabilitation Engineering 23 (1) (2015) 1–9. [17] A. R. Hassan, M. I. H. Bhuiyan, Automatic sleep scoring using statistical features in the emd domain and ensemble methods, Biocybernetics and

M

Biomedical Engineering 36 (1) (2016) 248 – 255.

ED

[18] F. Riaz, A. Hassan, S. Rehman, I. K. Niazi, K. Dremstrup, Emd-based temporal and spectral features for the classification of eeg signals using

PT

supervised learning, IEEE Transactions on Neural Systems and Rehabilitation Engineering 24 (1) (2016) 28–35.

CE

[19] O. Tsinalis, P. M. Matthews, Y. Guo, Automatic sleep stage scoring using time-frequency analysis and stacked sparse autoencoders, Annals

AC

of Biomedical Engineering 44 (5) (2016) 1587–1597.

[20] M. Peker, An efficient sleep scoring system based on eeg signal using complex-valued machine learning algorithms, Neurocomputing (2016) –.

37

ACCEPTED MANUSCRIPT

[21] G. Zhu, Y. Li, P. . Wen, Analysis and classification of sleep stages based on difference visibility graphs from a single-channel eeg signal, IEEE

CR IP T

Journal of Biomedical and Health Informatics 18 (6) (2014) 1813–1821. [22] S.-F. Liang, C.-E. Kuo, Y.-H. Hu, Y.-H. Pan, Y.-H. Wang, Automatic stage scoring of single-channel sleep eeg by using multiscale entropy and autoregressive models, Instrumentation and Measurement, IEEE

AN US

Transactions on 61 (6) (2012) 1649–1657.

[23] M. Ronzhina, O. Janouek, J. Kolrov, M. Novkov, P. Honzk, I. Provaznk, Sleep scoring using artificial neural networks, Sleep Medicine Reviews 16 (3) (2012) 251 – 263.

[24] A. R. Hassan, M. I. H. Bhuiyan, Automatic sleep stage classification,

M

in: Electrical Information and Communication Technology (EICT), 2013

ED

International Conference on, 2015, pp. 1–6. [25] B. Kemp, A. Zwinderman, B. Tuk, H. Kamphuisen, J. Oberye, Analysis of a sleep-dependent neuronal feedback loop: the slow-wave microconti-

PT

nuity of the eeg, Biomedical Engineering, IEEE Transactions on 47 (9)

CE

(2000) 1185–1194.

[26] C. Berthomier, X. Drouot, M. Herman-Stoca, P. Berthomier, J. Prado,

AC

D. Bokar-Thire, O. Benoit, J. Mattout, M.-P. dOrtho, Automatic analysis of single-channel sleep eeg: Validation in healthy individuals, Sleep 30 (11) (2007) 15871595.

[27] The dreams subjects database. URL http://www.tcts.fpms.ac.be/ devuyst/Databases/DatabaseSubjects/ 38

ACCEPTED MANUSCRIPT

[28] I. Selesnick, Wavelet transform with tunable Q-factor, Signal Processing, IEEE Transactions on 59 (8) (2011) 3560–3575.

CR IP T

[29] S. Patidar, R. B. Pachori, U. R. Acharya, Automated diagnosis of coro-

nary artery disease using tunable-Q wavelet transform applied on heart rate signals, Knowledge-Based Systems 82 (0) (2015) 1 – 10.

[30] A. R. Hassan, M. I. H. Bhuiyan, Computer-aided sleep staging using

AN US

complete ensemble empirical mode decomposition with adaptive noise and bootstrap aggregating, Biomedical Signal Processing and Control 24 (2016) 1 – 10.

[31] K. Fukunaga, Introduction to Statistical Pattern Recognition, 2nd Edi-

M

tion, 2009.

[32] C.-S. Huang, C.-L. Lin, L.-W. Ko, S.-Y. Liu, T.-P. Su, C.-T. Lin,

ED

Knowledge-based identification of sleep stages based on two forehead electroencephalogram channels, Frontiers in Neuroscience 8 (263).

PT

[33] A. Hassan, S. Bashar, M. Bhuiyan, On the classification of sleep states by means of statistical and spectral features from single channel electroen-

CE

cephalogram, in: Advances in Computing, Communications and Informatics (ICACCI), 2015 International Conference on, 2015, pp. 2238–

AC

2243.

[34] A. R. Hassan, M. A. Haque, Computer-aided gastrointestinal hemorrhage detection in wireless capsule endoscopy videos, Computer Methods and Programs in Biomedicine 122 (3) (2015) 341 – 353.

39

ACCEPTED MANUSCRIPT

[35] L. Breiman, Bagging predictors, Machine Learning 24 (2) (1996) 123– 140.

Systems Magazine, IEEE 6 (3) (2006) 21–45.

CR IP T

[36] R. Polikar, Ensemble based systems in decision making, Circuits and

[37] A. R. Hassan, Computer-aided obstructive sleep apnea detection using normal inverse gaussian parameters and adaptive boosting, Biomedical

AN US

Signal Processing and Control 29 (2016) 22 – 30.

[38] L. G. Doroshenkov, V. A. Konyshev, S. V. Selishchev, Classification of human sleep stages based on eeg processing using hidden markov models, BioMedical Engineering 41 (1) (2007) 24–28.

M

[39] S.-F. Liang, C.-E. Kuo, Y.-H. Hu, Y.-S. Cheng, A rule-based automatic sleep staging method, Journal of Neuroscience Methods 205 (1) (2012)

ED

169 – 176.

[40] C. Vural, M. Yildiz, Determination of sleep stage separation ability of

PT

features extracted from eeg signals using principal component analysis, Journal of Medical Systems 34 (2010) 83–89.

CE

[41] Y.-L. Hsu, Y.-T. Yang, J.-S. Wang, C.-Y. Hsu, Automatic sleep stage recurrent neural classifier using energy features of eeg signals, Neuro-

AC

computing 104 (0) (2013) 105 – 114.

[42] A. R. Hassan, S. K. Bashar, M. I. H. Bhuiyan, Automatic classification of sleep stages from single-channel electroencephalogram, in: India Conference (INDICON), 2015 Annual IEEE, 2015, pp. 1–6. 40

ACCEPTED MANUSCRIPT

[43] A. Iranzo, J. L. Molinuevo, J. Santamara, M. Serradell, M. J. Mart, F. Valldeoriola, E. Tolosa, Rapid-eye-movement sleep behaviour disorder

The Lancet Neurology 5 (7) (2006) 572 – 577.

CR IP T

as an early marker for a neurodegenerative disorder: a descriptive study,

[44] A. R. Hassan, M. I. H. Bhuiyan, A decision support system for automatic

sleep staging from eeg signals using tunable Q-factor wavelet transform

– 118.

AN US

and spectral features, Journal of Neuroscience Methods 271 (2016) 107

[45] R. G. Baraniuk, D. L. Jones, A signal-dependent time-frequency representation: optimal kernel design, IEEE Transactions on Signal Processing 41 (4) (1993) 1589–1602.

M

[46] D. L. Jones, R. G. Bariniuk, An adaptive optimal-kernel time-frequency representation, in: 1993 IEEE International Conference on Acoustics,

ED

Speech, and Signal Processing, Vol. 4, 1993, pp. 109–112 vol.4.

PT

[47] B. Boashash, S. Ouelha, An improved design of high-resolution quadratic time-frequency distributions for the analysis of nonstationary multicomponent signals using directional compact kernels, IEEE

AC

CE

Transactions on Signal Processing 65 (10) (2017) 2701–2713.

41