Most attempts at training computers for the difficult and time-consuming task of sleep stage classification involve a feature extraction step. Due to the complexity of multimodal sleep data, the size of the feature space can grow to the extent that it is also necessary to include a feature selection step. In this paper, we propose the use of an unsupervised feature learning architecture called deep belief nets (DBNs) and show how to apply it to sleep data in order to eliminate the use of handmade features. Using a postprocessing step of hidden Markov model (HMM) to accurately capture sleep stage switching, we compare our results to a feature-based approach. A study of anomaly detection with the application to home environment data collection is also presented. The results using raw data with a deep architecture, such as the DBN, were comparable to a feature-based approach when validated on clinical datasets. 1. Introduction One of the main challenges in sleep stage classification is to isolate features in multivariate time-series data which can be used to correctly identify and thereby automate the annotation process to generate sleep hypnograms. In the current absence of a set of universally applicable features, typically a two-stage process is required before training a sleep stage algorithm, namely, feature extraction and feature selection [1–9]. In other domains which share similar challenges, an alternative to using hand-tailored feature representations derived from expert knowledge is to apply unsupervised feature learning techniques, where the feature representations are learned from unlabeled data. This not only enables the discovery of new useful feature representations that a human expert might not be aware of, which in turn could lead to a better understanding of the sleep process and present a way of exploiting massive amounts of unlabeled data. Unsupervised feature learning and in particular deep learning [10–15] propose ways for training the weight matrices in each layer in an unsupervised fashion as a preprocessing step before training the whole network. This has proven to give good results in other areas such as vision tasks [10], object recognition [16], motion capture data [17], speech recognition [18], and bacteria identification [19]. This work presents a new approach to the automatic sleep staging problem. The main focus is to learn meaningful feature representations from unlabeled sleep data. A dataset of 25 subjects consisting of electroencephalography (EEG) of brain activity, electrooculography (EOG) of eye movements, and
References
[1]
K. ?u?mákováemail and A. Krakovská, “Discrimination ability of individual measures used in sleep stages classification,” Artificial Intelligence in Medicine, vol. 44, no. 3, pp. 261–277, 2008.
[2]
A. Flexer, G. Gruber, and G. Dorffner, “A reliable probabilistic sleep stager based on a single EEG signal,” Artificial Intelligence in Medicine, vol. 33, no. 3, pp. 199–207, 2005.
[3]
L. Johnson, A. Lubin, P. Naitoh, C. Nute, and M. Austin, “Spectral analysis of the EEG of dominant and non-dominant alpha subjects during waking and sleeping,” Electroencephalography and Clinical Neurophysiology, vol. 26, no. 4, pp. 361–370, 1969.
[4]
J. Pardey, S. Roberts, L. Tarassenko, and J. Stradling, “A new approach to the analysis of the human sleep/wakefulness continuum,” Journal of Sleep Research, vol. 5, no. 4, pp. 201–210, 1996.
[5]
N. Schaltenbrand, R. Lengelle, M. Toussaint et al., “Sleep stage scoring using the neural network model: comparison between visual and automatic analysis in normal subjects and patients,” Sleep, vol. 19, no. 1, pp. 26–35, 1996.
[6]
H. G. Jo, J. Y. Park, C. K. Lee, S. K. An, and S. K. Yoo, “Genetic fuzzy classifier for sleep stage identification,” Computers in Biology and Medicine, vol. 40, no. 7, pp. 629–634, 2010.
[7]
L. Zoubek, S. Charbonnier, S. Lesecq, A. Buguet, and F. Chapotot, “A two-steps sleep/wake stages classifier taking into account artefacts in the polysomnographic signa,” in Proceedings of the 17th World Congress, International Federation of Automatic Control (IFAC '08), July 2008.
[8]
G. Luo and W. Min, “Subject-adaptive real-time sleep stage classification based on conditional random field,” in Proceedings of the American Medical Informatics Association Annual Symposium (AMIA '07), pp. 488–492, 2007.
[9]
T. Penzel, K. Kesper, V. Gross, H. F. Becker, and C. Vogelmeier, “Problems in automatic sleep scoring applied to sleep apnea,” in Proceedings of the 25th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (IEEE EMBS '03), pp. 358–361, September 2003.
[10]
G. E. Hinton, S. Osindero, and Y. W. Teh, “A fast learning algorithm for deep belief nets,” Neural Computation, vol. 18, no. 7, pp. 1527–1554, 2006.
[11]
Y. Bengio, P. Lamblin, D. Popovici, and H. Larochelle, “Greedy layer-wise training of deep networks,” in Advances in Neural Information Processing Systems (NIPS '06), vol. 19, pp. 153–160, 2006.
[12]
M. Ranzato, C. Poultney, S. Chopra, and Y. LeCun, “Efficient learning of sparse representations with an energy-based model,” in Proceedings of the Advances in Neural Information Processing Systems (NIPS '06), J. Platt, T. Hoffman, and B. Sch?lkopf, Eds., MIT Press, 2006.
[13]
Y. Bengio and Y. LeCun, “Scaling learning algorithms towards AI,” in Large-Scale Kernel Machines, L. Bottou, O. Chapelle, D. DeCoste, and J. Weston, Eds., MIT Press, 2007.
[14]
Y. Bengio, “Learning deep architectures for AI,” Tech. Rep. 1312, Department of IRO, Universite de Montreal, 2007.
[15]
I. Arel, D. Rose, and T. Karnowski, “Deep machine learning—a new frontier in artificial intelligence research,” IEEE Computational Intelligence Magazine, vol. 14, pp. 12–18, 2010.
[16]
V. Nair and G. E. Hinton, “3-d object recognition with deep belief nets,” in Proceedings of the Advances in Neural Information Processing Systems (NIPS '06), 2006.
[17]
G. Taylor, G. E. Hinton, and S. Roweis, “Modeling human motion using binary latent variables,” in Proceedings of the Advances in Neural Information Processing Systems, 2007.
[18]
N. Jaitly and G. E. Hinton, “Learning a better representation of speech sound waves using restricted boltzmann machines,” in Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP '11), 2011.
[19]
M. L?ngkvist and A. Loutfi, “Unsupervised feature learning for electronic nose data applied to bacteria identification in blood,” in NIPS Workshop on Deep Learning and Unsupervised Feature Learning, 2011.
[20]
D. F. Wulsin, J. R. Gupta, R. Mani, J. A. Blanco, and B. Litt, “Modeling electroencephalography waveforms with semi-supervised deep belief nets: fast classification and anomaly measurement,” Journal of Neural Engineering, vol. 8, no. 3, Article ID 036015, 2011.
[21]
G. E. Hinton, “Training products of experts by minimizing contrastive divergence,” Neural Computation, vol. 14, no. 8, pp. 1771–1800, 2002.
[22]
A. Rechtschaffen and A. Kales, A Manual of Standardized Terminology, Techniques and Scoring System for Sleep Stages of Human Subjects, U.S. Government Printing Office, Washington DC, USA, 1968.
[23]
M. Hirshkowitz, “Standing on the shoulders of giants: the Standardized Sleep Manual after 30 years,” Sleep Medicine Reviews, vol. 4, no. 2, pp. 169–179, 2000.
[24]
S. L. Himanen and J. Hasan, “Limitations of Rechtschaffen and Kales,” Sleep Medicine Reviews, vol. 4, no. 2, pp. 149–167, 2000.
[25]
L. R. Rabiner and B. H. Juang, “An introduction to hidden markov models,” IEEE ASSP Magazine, vol. 3, no. 1, pp. 4–16, 1986.
[26]
G. E. Hinton, A Practical Guide to Training Restricted Boltzmann Machines, 2010.
[27]
S. Charbonnier, L. Zoubek, S. Lesecq, and F. Chapotot, “Self-evaluated automatic classifier as a decision-support tool for sleep/wake staging,” Computers in Biology and Medicine, vol. 41, no. 6, pp. 380–389, 2011.
[28]
A. Schl?gl, C. Keinrath, D. Zimmermann, R. Scherer, R. Leeb, and G. Pfurtscheller, “A fully automated correction method of EOG artifacts in EEG recordings,” Clinical Neurophysiology, vol. 118, no. 1, pp. 98–104, 2007.
[29]
A. L. Goldberger, L. A. Amaral, L. Glass et al., “PhysioBank, PhysioToolkit, and PhysioNet: components of a new research resource for complex physiologic signals,” Circulation, vol. 101, no. 23, pp. E215–220, 2000.
[30]
L. Zoubek, S. Charbonnier, S. Lesecq, A. Buguet, and F. Chapotot, “Feature selection for sleep/wake stages classification using data driven methods,” Biomedical Signal Processing and Control, vol. 2, no. 3, pp. 171–179, 2007.
[31]
G. Huang, H. Lee, and E. Learned-Miller, “Learning hierarchical representations for face verification with convolutional deep belief networks,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR '12), 2012.
[32]
D. Yu, L. Deng, I. Jang, P. Kudumakis, M. Sandler, and K. Kang, “Deep learning and its applications to signal and information processing,” IEEE Signal Processing Magazine, vol. 28, no. 1, pp. 145–154, 2011.
[33]
M. Ranzato, A. Krizhevsky, and G. E. Hinton, “Factored 3-way restricted boltzmann machines for modeling natural images,” in Proceedings of the 30th International Conference on Artificial Intelligence and Statistics, 2010.
[34]
J. Ngiam, A. Khosla, M. Kim, H. Lee, and A. Y. Ng, “Multimodal deep learning,” in Proceedings of the 28th International Conference on Machine Learning, 2011.
[35]
A. R. Osborne and A. Provenzale, “Finite correlation dimension for stochastic systems with power-law spectra,” Physica D, vol. 35, no. 3, pp. 357–381, 1989.
[36]
E. Pereda, A. Gamundi, R. Rial, and J. González, “Non-linear behaviour of human EEG: fractal exponent versus correlation dimension in awake and sleep stages,” Neuroscience Letters, vol. 250, no. 2, pp. 91–94, 1998.
[37]
T. Gasser, P. Baecher, and J. Moecks, “Transformations towards the normal distribution of broad band spectral parameters of the EEG,” Electroencephalography and Clinical Neurophysiology, vol. 53, no. 1, pp. 119–124, 1982.