In 2004 a reality television program called ‘Shattered’ was broadcast on Channel 4 in the United Kingdom. In the show, a number of contestants struggled to stay awake for as long as possible. They remained awake for days on end (the winner managed to stay awake for 178 h) but at a considerable cost. Lack of sleep resulted in a severe loss in concentration capacity and extreme mood swings. At some point, the contestants began to behave in an erratic fashion and they even started to hallucinate and to have delusional thoughts (known as hypnagogia). These phenomena disappeared after one or two good nights’ sleep. It goes without saying that such dramatic behavioral effects reflect changes in functioning of the sleep-deprived nervous system.

Sleep loss can have various origins with pathological causes as first to mention. In this context, one discriminates between sleep disorders (dyssomnias) as such, like obstructive sleep apnoea, and other disorders that involve sleep loss, such as clinical depression, chronic pain, and restless-legs syndrome. Sleep loss can also be induced in test animals or human volunteers to investigate the neurobehavioral consequences of insufficient sleep. Moreover, as humans, we can choose to voluntarily adopt a pattern of waking and sleeping that opposes the body’s urge to rest, e.g., extensive traveling across time zones (chronic jet lag) and working on night shifts, which may lead to chronic sleep loss [1]. Chronic sleep loss, or sleep debt [2], is becoming increasingly common and affects millions of people, especially in more developed countries where sleep debt is probably due to the increase in working hours in certain professions, such as medicine [3].

In the literature on the neurobehavioral effects of sleep loss a distinction is made between sleep reduction, sleep fragmentation, and total sleep loss or sleep deprivation [4]. Sleep reduction refers to a situation in which an organism falls asleep later than usual or wakes up earlier than dictated by its physiological needs, while sleep fragmentation refers to the case in which an organism is permitted to sleep but wakes up and falls asleep again at certain intervals. In general, sleep fragmentation constitutes a clinical symptom, but it may also be induced experimentally in healthy subjects. Finally, sleep deprivation (SD) can be considered an extreme case of sleep reduction, in that the organism is simply awake for a prolonged period of time, as for example when one or more nights of sleep are skipped. The distinction between sleep reduction, sleep fragmentation, and SD serves mainly as a useful heuristic and is not assumed to imply qualitatively different neurobehavioral effects. For example, Bonnet and Arand [5] demonstrated that all these forms of sleep loss may lead to very similar effects on the experience of sleepiness, psychomotor performance, and hormonal disturbances.

In the present article, we will review part of the vast experimental literature on sleep loss involving human participants, in combination with a few selected animal studies. In particular, we will focus on the behavioral and neural effects of SD with the aim to elucidate the neurophysiological underpinnings of SD-induced deterioration in cognitive functioning and to identify directions for future research. To this end, we review the behavioral and neural effects of SD according to three levels of observation, called the macro-, meso- and micro-level [610]. With the term macro-level, we refer to macroscopic behaviors including cognition, emotional processes, muscle activity, kinematics, and motion, as well as collective activities of large, crudely defined brain structures, e.g., prefrontal cortex (PFC), thalamus, and hippocampus. The growing field of cognitive neuroscience can be considered an attempt to link macroscopic behavior with large-scale neural activity. Modern neuroimaging techniques like functional magnetic resonance imaging (fMRI) and electro- and/or magneto-encephalography (EEG and MEG) exploit the neurons’ overall metabolism or record synchronized firing behavior of neural assemblies and allow for zooming in on neural activity and ensembles at the meso-level. With this term, we refer to neural activity within large brain areas down to ensembles of a smaller number of cells. That is, ensembles at the meso-level still form more or less well-defined functional units, both in terms of structure, such as a hyper-column in the visual cortex [11], and activity, i.e., neural synchronization [see e.g., refs. 12, 13]. Finally, the micro-level refers to the molecular and cellular level, that is, the level of ion channels, gene products and protein synthesis. This is the level at which very fast chemical processes occur that constitute the physical or chemical building blocks of brain function. Even though the boundaries between macro-, meso- and micro-levels are not clear cut, they provide a useful heuristic for discussing effects of SD and linking SD phenomena across levels. Although we will review effects of SD on all scales, the emphasis will be on neural activity at the macro-level and at the meso-level, that is, on behavior and interactions among comparatively large brain structures. Where possible, we will establish a link between the macro- and/or meso-level and the micro-level, focusing on effects of SD on cortical activity in relation to impairment in cognitive performance. We will first establish a link between the behavioral effects of SD and its effects on large-scale brain activity of SD. Subsequently, we will show how effects of SD on behavior and brain activity (macro- and meso-levels) are matched by changes in cellular and molecular changes (micro-level) as reported in animal studies.

SD affects behavior and coarse-grained brain activity — the macro-level

Applied and fundamental research has identified a broad variety of detrimental consequences of SD on performance. In more applied fields, there is a major interest in groups of workers who customarily do not receive enough sleep [1419]. For example, a study involving doctors working in a hospital reported that, after a 32-h shift, the resulting lack of sleep had a profound negative effect on mood and cognitive performance [3]. However, several studies also reported beneficial effects of SD, e.g., one night of SD can have beneficial effects on mood in the majority of patients diagnosed with depression, possibly through influences on the circadian rhythm and the arousal system [2022]. Unfortunately, the depressive symptoms usually re-appear after a recovery sleep [23, 24]. In addition, epidemiological studies have shown that mild sleep restriction decreases mortality in long sleepers, putatively because sleep restriction might resemble dietary restriction as a potential aid to survival [25]. Moreover, spending excessive time in bed can elicit daytime lethargy and exacerbate sleep fragmentation, resulting in a vicious cycle of further time in bed and further sleep fragmentation. However, these beneficial effects of mild sleep restriction reverse at sleep durations of less than 6.5 h per night [25].

Mood, memory, and cognition

Fundamental research has shown that SD profoundly affects human functioning and that its effects are so widespread that it is difficult to find functions that remain untouched [2628]. The literature further suggests that consolidation of newly learned material (e.g., novel words or novel perceptual skills) is disrupted after loss of a night’s sleep [29, 30]. Apart from affecting performance, sleep loss invariably leads to a strong subjective feeling of sleepiness, negative mood, and stress [28]. In extreme cases, SD might even lead to hallucinations, and test animals that are sleep-deprived for an indefinite period of time will eventually die. It remains unclear, however, whether the detrimental effects of SD on performance are simply due to a lack of sleep as such or to the accompanying stress [cf. ref. 31]. Stress typically results in a rise in cortisol levels, which, for example, led McEwen [32] to suggest that chronic SD acts as a chronic stressor that yields allostatic load on the body, which, in turn, negatively affects the immune and the nervous system. A study by Ruskin and colleagues [33], however, produced opposing results, in that both adrenolectomized rats, i.e., rats with a blocked adrenal stress system, and intact rats suffered to an equal degree from SD.

Recent studies have shown impaired performance of executive functioning including measures of verbal fluency, creativity, planning skills [4, 34], novelty processing [35], and driving performance [36]. The impact of SD is likely to be particularly prominent in tasks that strongly depend on attention, i.e., tasks that require other than well-learned automatic responses will be most vulnerable [37]. Besides the vast literature on SD-related performance decrements in novel tasks, there are also several studies showing that SD leads to performance deficits in vigilance tasks [28, 3840]. These behavioral effects point to the involvement of brain structures that have been associated with attention and arousal. In particular, several studies have hinted at a central role for the PFC in relation to SD [4, 41, 42]. The PFC is a neocortical region that is known to support a diverse and flexible repertoire of behavioral functions and is most elaborated in primates. It consists of a massive network, connecting perceptual, motor, and limbic regions within the brain, and is important whenever top-down processing is needed, i.e., during executive functions and attention [43, 44]. Impaired attention and cognitive performance due to SD therefore suggest decreased brain activity and function in the PFC [45]. Similarly, the disrupted memory consolidation after SD indicates involvement of the hippocampus, a crucial structure involved in learning and the consolidation of newly learned material [46, 47]. Indeed, Hairston and colleagues [48] recently found that sleep restriction in the rat impaired hippocampus-dependent learning.

Staying awake

Another way to study the neural effects of sleep is by examining agents that promote wakefulness. The subjective feelings of sleepiness and the objective decrements in psychomotor functioning can be temporarily reversed by administering psychostimulants, such as caffeine. Caffeine is the most widely consumed stimulant in the world [49] and has been shown to produce significant alerting and long-lasting beneficial mood effects in sleep-deprived individuals [50]. Likewise, the military is actively seeking ways to prolong wakefulness in soldiers without hampering their performance in sustained operations. As a case in point, Baranski and colleagues [51] studied effects of the psychostimulant modafinil on mood and performance in a group of sleep-deprived soldiers. Interestingly, a dose of about 300 mg modafinil resulted in a level of cognitive performance that was virtually identical to baseline. Wesensten and colleagues [52] directly compared effects of caffeine, modafinil, and amphetamine on a group of healthy volunteers who had remained awake for 85 h. It was found that all agents improved vigilance, and that caffeine and modafinil improved executive functioning. For a review, describing in more detail the neuropharmacology of caffeine and other stimulants, see Boutrel and Koob [53]. In brief, caffeine seems to exert its effect on wakefulness via two complementary routes; on the one hand, it activates cholinergic neurons in the forebrain, while on the other, it inhibits sleep-promoting neurons in the hypothalamus. A behavioral method to promote wakefulness is to engage in challenging tasks that require a high degree of motivation [37]. Interestingly, even body posture affects the tendency to fall asleep: Caldwell and colleagues [54] found that sleep-deprived participants were better able to stay awake in a standing upright posture than when sitting, as demostrated by psycho-motor performance and EEG arousal.

SD affects more fine-grained brain activity — the meso-level

Over the years, a wide range of non-invasive methods has become available to study human brain activity [see e.g., ref. 55]. Broadly speaking, these methods fall in two categories. The first class of methods serves to accurately identify locations in the brain and is based on the premise that increased activity and recruitment of neural tissue is accompanied by an increased metabolism in terms of oxygen and/or glucose consumption. Well-known methods within this class are positron emission tomography (PET) and fMRI. While PET measures hemodynamic changes by marking blood with a radioactive tracer, fMRI measures the blood oxygenation level-dependent (BOLD) contrast [5658]. As blood flow to specific brain region increases with increased neural activity, both techniques provide indirect measures of this activity. To extract the brain activity related to a specific cognitive function, one typically determines the difference in brain activity under baseline and activation conditions [5860].

The second class of methods serves to study the temporal dynamics of neural activity. Well-known methods within this class are EEG and MEG. With EEG, one measures voltage or potential differences between different parts of the brain using electrodes placed on the scalp. With MEG, in contrast, one measures magnetic fields that derive from the flow of ionic currents in the (cortical) neuron dendrites [61]. In analyzing encephalographic data, identification of event-related potentials (ERPs in the case of EEG) or event-related fields (ERFs in the case of MEG) is a common technique. By averaging traces over multiple trials that are aligned at a specific event (usually, the onset of an external stimulus), brain activity related to cortical processing can be extracted. That is, one identifies certain time- or phase-locked components and omits unwanted ‘noise’, i.e., background activity that is not phase-locked to the event in question [6264], yielding a good signal-to-noise ratio. A range of ERP and ERF components have been identified that reflect specific forms of information processing, related to sensory, motor, and/or cognitive functions, as well as specific subject characteristics, like age and health status [see e.g., refs. 65, 66].

Event-related components

With respect to SD, two characteristic ERP/ERF components have received special attention in the literature, namely N1 and P300 (sometimes called P3). N1 represents a negative polarity deflection in the waveform about 100 ms after stimulus presentation. It is associated with the processing of auditory, visual, or tactile stimuli in primary sensory areas [cf. ref. 67]. However, N1 is not a purely sensory phenomenon but is affected by attention, e.g., the amplitude of the auditory N1 can be increased by simply asking participants to attend to one ear [68, 69] or by administering a small dose of caffeine [70, 71]. The importance of N1 as a marker for attentional processing is underscored by the clinical observation that patients with frontal lesions exhibit a reduction in N1 amplitude, and a concomitant disruption of attentional capacities [72]. Similarly, N1 amplitude reduction can also be induced by SD. For example, using an auditory oddball paradigm (in which participants had to respond to infrequently appearing target stimuli), Cote and colleagues [73] observed a marked N1 amplitude reduction in participants who had suffered a night of sleep fragmentation. N1 amplitude reduction after SD is not restricted to auditory responses [7476], but is also observed for visual- [77] and motor-evoked potentials [76]. Figure 1 shows the reduction in the auditory- and motor-related field following SD. These experimental findings point to reduced responsiveness of sensory areas to peripheral input after SD due to reduced attention [76]. Similarly, the increased N1 amplitude after a small dose of caffeine has also been associated with attention [70, 71].

Figure 1
figure 1

Reduction of the auditory- (upper panel) and motor-related (lower panel) response after SD. Left: spatial distribution of both responses showing event-related activity above either bilateral auditory areas or the left sensorimotor area (head is viewed from above using a Mercator-mapping; nose is upwards). Middle: the strength of event-related components in the control (blue) and sleep-deprived (red) condition. Right: the corresponding event-related fields [see ref. 76, for more details].

In contrast to N1, which is localized above sensory cortices, P300 appears as a positive deflection of the EEG voltage at the frontal scalp around 300 ms post-stimulus. P300 is associated with frontal lobe functioning and can be elicited by presenting the participant with a visual or auditory stimulus that is either unexpected or highly relevant for the task at hand [78]. Following SD, the onset of P300 is somewhat delayed and its amplitude reduced [4]. More recently, it has been suggested that two subtypes of P300 should be distinguished, a so-called ‘novel P3’ and a ‘target P3’ [35]. The novel P3 is thought to originate from the anterior cingulate cortex or the supplementary motor area and to be related to the detection of unexpected stimuli. The target P3, in contrast, is thought to reflect neural activity at the temporal-parietal junction and to be related to the detection of anticipated stimuli [79]. Consistent with this idea, Gosselin and colleagues [35] found that both novel and target P3 were reduced in amplitude after 36 hours of waking, and that performance on attention-demanding tasks was deteriorated. The authors concluded that SD affects a whole attentional network, consisting of several interconnected cortical regions.

As was the case for N1, caffeine affects P300 in a manner opposite to that of SD. Deslandes and colleagues [80] found in an oddball paradigm that administration of a 400-mg dose of caffeine reduced the latency and increased the amplitude of P300. Given the current collection of empirical observations, it seems reasonable to conclude that the amplitude reduction of N1 and P300 after SD and the amplitude enhancement of those components after a dose of caffeine are both related to changes in attention and thus to changes in sensory processing, albeit in opposite directions. This conclusion is further supported by the observation that both N1 and P300 are influenced by noradrenergic, dopaminergic, and cholinergic systems, i.e., systems which are known to mediate attention and arousal [81, 82].

Frequency contents of the encephalogram

Encephalographic patterns can also be characterized by the amplitude and frequency of electric (or magnetic) activity. The normal human EEG and MEG show amplitudes of 20–100 µV and a few hundred fT, respectively, and frequencies up to about 100 Hz. As explained, time-dependent changes of the neural activity can be addressed via event-related studies, whereas spectral analysis allows for investigating ongoing cortical dynamics, that is, activity that is not necessarily event-locked. The latter approach has a long tradition in the study of various stages of sleep and wakefulness as well as pathophysiological situations such as seizures. Traditionally, encephalographic frequencies are divided into several frequency bands: delta (1–4 Hz), theta (4–8 Hz), alpha (8–12 Hz), beta (13–30 Hz), and gamma (30–90 Hz). During sleep characteristic changes occur in the spectral power within those frequency bands [see e.g., refs. 8385], which have been instrumental in the identification and subsequent definition of distinct sleep stages [86]. For instance, so-called slow-wave sleep (sleep stages III and IV) is characterized by strong delta activity and low responsiveness to external stimuli and might have a recovery function [87].

For sleep-deprived participants, the occurrence of slow-wave sleep is enhanced on the first recovery night after SD, whereas REM sleep, which is signified by strong desynchronization of the encephalogram, does not differ from the baseline value [84, 8891]. In the waking EEG, a reduction in alpha activity after SD was already observed by Blake and Gerard [92] and has been replicated many times [for an overview, see ref. 31]. More recently, a progressive increase in theta activity during prolonged wakefulness has been reported [93, 94]. In general, the waking EEG appears to undergo changes in several frequency bands that might be attributed to circadian as well as homeostatic processes. The distinct variations in the theta and alpha band may represent electrophysiological correlates of different aspects of a circadian change in arousal [9597]. Moreover, the combined increase in theta activity and decrease in alpha activity seems to correspond to a slowing of encephalographic rhythms, that is, a shift of spectral power toward lower frequencies. A prevalence of slow-wave activity might thus provide an index of sleep propensity and/or cortical deactivation [84, 91, 94, 98].

Spatial patterns of the encephalogram

Besides amplitude and frequency characteristics, SD affects the spatial distribution of encephalographic activity over the scalp. Several studies reported an increase in the degree of correlation between activity in both hemispheres as a result of prolonged wakeful-ness [see e.g., ref. 99, 100]. Furthermore, an anterior shift of activity has been reported after 24 h of wakefulness: the overall power, as well as the power contained in the alpha band, decreased at occipital and temporal channels and increased at central and frontal channels [76]. These results are consistent with encephalographic studies on the sleep-wake cycle that revealed an anterior shift of activity during the wakefulness-sleep transition [101103]. The changes in encephalographic activity along the anterior-posterior axis seem to indicate that effects of SD are local and mainly pertain to prefrontal areas [104]. That is, increased synchronization of cortical activity, resulting in an increased amplitude of encephalographic activity, is thought of as a sign of cortical deactivation or cortical idling [63]. Hence, these data seem in line with fMRI results showing reduced activity in the PFC. Note, however, that the change in spatial distribution cannot be simply attributed to a single cortical structure, not least because of the low spatial resolution of encephalographic data.

Neuroimaging studies

The detrimental effect of SD on cognitive performance has instigated research aimed at pinpointing influences on the PFC using neuroimaging methods like PET and fMRI. As already indicated above, a number of studies have observed reduced neural activity in the dorsolateral PFC (see Fig. 2) following SD, and a subsequent decrement in cognitive performance [105108]. Findings of decreased regional brain activity suggest that neurons cannot keep pace with high task load requirements in the case of complex task performance and/or sustained task performance during SD [108]. Other studies, most notably by the group of Drummond, however, found an opposite pattern of results, namely an increase in bilateral dorsolateral PFC activity induced by SD [105108]. For example, in a verbal learning task, Drummond and colleagues [105] found increased brain activity in the dorsolateral PFC and parietal cortex after 35 h of wakefulness, although subsequent free recall was significantly impaired after SD. They concluded, therefore, that, depending on task difficulty, there can be dynamic, compensatory changes in neural recruitment and, furthermore, that those changes bear a striking resemblance to neural changes in normal aging. In support, Choo and colleagues [109] found that the increase in activity in the PFC after SD was dependent on working memory load. However, they concluded that the mixed results obtained from imaging studies following SD suggest that studying the neural correlates of working memory in this setting is more complex than originally envisioned.

Figure 2
figure 2

Illustration of the lateral part (Brodmann’s areas 45 and 46 and the lateral parts of areas 9 to 12) and the ventromedial part of the PFC (the medial parts of areas 10 and 11). The dorsolateral PFC is engaged in executive functions such as working memory, whereas the ventromedial PFC is part of the limbic system involved in decision making and social cognition [42]. Figure adopted from Gazzaniga et al. [55].

Thus, there is still debate as to whether the neural activity in PFC increases or decreases after SD. In part, the discrepancy in the empirical findings regarding PFC activity might be caused by methodological differences, like the experimental task of choice, the amount of SD, the age of the participants, and the method used to analyze the data. Although all these factors affect PFC activity, its central role in SD and sleep is now well established [4, 42, 104]. As discussed above, the PFC is thought to play an important role in the higher-order cognitive impairments noted in various SD experiments. Thomas and coworkers [45] reported, besides a global decrease in brain activity, a decrease in activity throughout the thalamus and PFC, both of which are part of the network mediating alertness, attention and higher-order cognitive processes [110].

The prominence of the PFC in attentional processes has received considerable support and provides further insight into the functional role of the attentional network during SD [45, 111, 112]. In particular, the interaction between bottom-up arousal and top-down attention, as regulated through PFC, appears crucial [111, 112]. In this context, it is interesting to note that the PFC has also a top-down influence on the basal forebrain (BF) cholinergic system. The cholinergic system is part of the ascending activation system projecting to all cortical areas, but it only receives cortical input via the limbic system [110]. Hence, only the limbic system can effectively induce a rapid change of activity in these nuclei, and activation of these pathways provides a mechanism for modulating the neural responses to novel and motivationally relevant events, facilitating further processing of these events [110, 113, 114].

SD-related cellular and molecular changes — the micro-level

A neuromodulator that has been studied extensively in sleep research is adenosine (AD). AD is a purinergic messenger that regulates many physiological processes, particularly in excitable tissues such as heart and brain. It helps to couple the rate of energy expenditure to the energy supply by either reducing the activity of the tissue or by increasing the energy supply to the tissue. AD plays a variety of roles in the brain as intercellular messenger and has been shown to be involved in processes such as sleep regulation and arousal [115]. As a neuromodulator, AD has a general inhibitory effect on the release of most other neurotransmitters, such as the excitatory glutamate and acetylcholine [116118], and thereby reduces excitability throughout the brain. Indeed, systemic or intracerebroventricular administration of AD has been shown to promote sleep and decrease the waking state [119]. The role of AD in sleep regulation has been supported further by studies showing a progressive increase of extracellular AD in the BF during prolonged wakefulness and a decrease during subsequent recovery sleep [120123]. In addition, pharmacological manipulations involving AD receptors have shown that agonists generally promote sleep [124], whereas antagonists such as caffeine reduce sleep [125]. Hence, there is ample evidence that AD plays a central role as endogenous sleep factor [cf. refs. 118, 126].

Effects of SD on AD levels

Several mechanisms are capable of influencing the extracellular AD level, such as the modulation of AD anabolic and catabolic enzyme activity and AD transport rate through the cell membrane. Because of the presence of equilibrative transporters, regulation of intracellular AD concentrations is critical for the regulation of extracellular AD [115, 127]. Intracellular AD plays an important role in the energy metabolism of the cell and AD concentrations increase with cellular metabolism [128]. Hence, the increase in AD levels during waking is assumed to be caused by an increase in metabolic activity during waking which is reversed again during sleep [120, 129, 130]. The increase in extracellular AD is particularly manifest in the BF, where the concentration progressively increases with increased duration of waking to reach about twice that measured at the onset of the experiment [120]. BF is part of the arousal system of the brain and consists of cholinergic neurons that have long-range projections and are active during wake and REM sleep [131]. As a result, Strecker and colleagues [118] suggested that during prolonged wakefulness, extracellular AD accumulates selectively in the BF and promotes the transition from wakefulness to sleep by inhibiting cholinergic and non-cholinergic BF neurons. Indeed, in the BF it has been demonstrated that AD inhibits both cholinergic neurons and a subset of non-cholinergic neurons [132]. The exclusive role of the cholinergic BF neurons as transducers of sleep drive, however, has been challenged by a study on rats that were administered 192-IgG-saporin, resulting in a 95% lesion of the BF cholinergic neurons. Because these rats had intact sleep homeostasis, these results indicate that neither the activity of the BF cholinergic neurons nor the accumulation of AD in the BF during wakefulness is necessary for sleep drive [133].

Extracellular AD exerts its inhibitory effect through different AD receptors. Although there are four subtypes of AD receptors, A1, A2A, A2B, and A3, several lines of evidence indicate that only the A1 and A2A receptors are involved in mediating the sleep-inducing effect of AD [115, 127, 134]. In support, caffeine acts as an AD receptor antagonist, binds to the A1 and A2A receptors [49], but exerts its arousal effect mainly via the latter [135]. Again, however, the lack of AD A1 receptors does not prevent the homeostatic regulation of sleep [136], indicating that effects of AD are not restricted to either a single cell or receptor type, but are more general.

Effects of AD and acetylcholine on macroscopic brain activity

As also put forward by Strecker and colleagues [118], the build-up of AD during SD might play a central role in bringing about the effects of SD on macroscopic brain activity by inhibiting the ascending activating nuclei important for attention and arousal. In particular, the inhibition of the cholinergic nuclei results in reduced cortical levels of (ACh), which have been shown to result in similar macroscopic effects as SD. Although some of these studies were not carried out for the purpose of investigating the effects of SD, we will discuss part of this literature showing that the decrease in ACh and the increase in AD results in similar effects as SD. That is, the inhibitory effect of AD extends to the cholinergic neurons of the mesopontine tegmentum, where it also reduces the excitability of those neurons [137]. The cholinergic neurons form a continuous neural network extending from the basal ganglia and BF of the telencephalon to the ventral horn of the spinal cord and since the cholinergic nuclei have a widespread influence throughout the central nervous system, they are particularly capable of affecting macroscopic brain activity (see Fig. 3) [138].

Figure 3
figure 3

Simplified diagram of the ascending activating system in the human brain. Adopted from http://web.lemoyne.edu/~hevern/psy340/lectures/; see text for further details.

First, the cholinergic neurons have a profound effect on thalamocortical arousal [81, 138141]. Electrical stimulation of the cholinergic neurons or direct application of ACh results in prolonged depolarization of thalamocortical neurons affecting the behavioral state of arousal [142, 143]. The thalamocortical network plays a central role in the generation of cortical rhythms, i.e., the progressive hyperpolarization of the thalamocortical cells during sleep onset causes a shift toward large-amplitude, low-frequency oscillations, as measured with EEG and MEG [139]. Thus, the build-up of AD during SD could modulate brain oscillations indirectly by suppressing other neurotransmitter systems such as the cholinergic nuclei, and ACh may thus act as an endogenous mediator of SD-induced increases in slow-wave activity [130, 144, 145]. In addition, ACh can also affect cortical rhythms directly. For example, Szerb [146] showed in anesthetized cats that stimulation of the reticular formation produced increased cortical ACh and reduced cortical low-frequency activity as measured with EEG.

Second, reduced cortical ACh levels result in the attention-related effects on the amplitude of ERP/ERF components, such as the N1 and P300. The most common effect of ACh is an enhancement of the neural discharge evoked by sensory stimuli. Consequently, the cholinergic system modifies sensory processing in the cerebral cortex by modulating the effectiveness of afferent inputs to cortical neurons, which suggests that cortical ACh modulates the general efficacy of the cortical processing of sensory or associational information [141, 147, 148]. The effects of SD on N1 and P300 could thus be an effect of reduced cortical ACh levels caused by inhibition of the cholinergic neurons by AD. In support, animal research indicates that the middle-latency auditory-evoked potential ‘wave A’, which is the equivalent of the N1 in humans, was reduced in direct proportion to the number of damaged cholinergic cells [149]. Similarly, the ‘cat P300’ disappeared after septal lesions, and the lesioned cats were characterized by marked ACh depletion [150]. It therefore seems that cholinergic systems play an important, albeit not exclusive, role in the mediation of P300 potentials in cats [151]. Furthermore, the novelty-related P300 response can be abolished by cholinergic antagonists in human participants [152]. These data support the notion that cortical cholinergic inputs mediate diverse attentional functions and may be related to the attentional effects reported after SD [114, 141].

Moreover, the increased involvement of the PFC during SD, as reported in several studies [105108], might be caused by deteriorated performance of cortical areas responsible for perceptual processing. For example, Furey and colleagues [153] found that ACh increased the selectivity of neural responses in extrastriate cortices during visual working memory, particularly during encoding, and increased the participation of ventral extrastriate cortex during memory maintenance. In parallel, they found decreased activation of the anterior PFC. These results indicate that cholinergic enhancement improves memory performance by augmenting the selectivity of perceptual processing during encoding, thereby simplifying processing demands during memory maintenance and reducing the need for prefrontal participation. Likewise, a reduced ACh level during SD would thus deteriorate perceptual processing and thereby place a higher burden on PFC, consistent with the notion of the compensatory role of PFC during SD [108].

Third, increased cortical AD levels also directly suppress cortical responses to incoming stimuli. Several studies have demonstrated that AD applied to the somatosensory cortex reduces the early components of the potentials evoked by peripheral sensory stimuli [154, 155] indicating that thalamocortical excitation is modulated by AD receptors. AD equally reduces thalamocortical inputs onto both inhibitory and excitatory neurons, suggesting that increased cortical AD levels result in a global reduction in the impact of incoming sensory information and thereby reduce the influence of sensory inputs on cortical activity (at least, in rodents) [156]. Again, these effects of AD on cortical activity are consistent with effects of SD on event-related encephalographic activity, i.e., both point at reduced effects of peripheral sensory stimuli on the cerebral cortex [cf. ref. 76]. In sum, the effects of increased AD and decreased ACh levels match the macroscopic effects of SD, such as the reduced ERP components and changes in EEG rhythms. Furthermore, the role of AD during SD has been established by various studies and shows how, by reducing excitability throughout the brain, the increase in AD may cause effects of SD as observed in brain activity on a macroscopic level.

Summary and final remarks

Both in terms of cause and effect, SD is a multifaceted phenomenon. Sleep loss, and the associated deterioration in performance, affects a large portion of the population and is thus of great social and economical importance [157]. On a macroscopic level, SD has widespread detrimental effects, influencing mood, memory, and cognitive performance. Most cognitive research agrees that SD mainly affects executive functions and is particularly manifest in novel tasks, while automated tasks are hardly affected. This readily hints at the involvement of neural structures related to attentional processes, most notably the PFC. The reduced amplitude of the N1 and P300 component in encephalographic recordings after SD mirrors effects of reduced attention and points to reduced cortical responsiveness to incoming stimuli. Neuroimaging studies (PET and fMRI) revealed effects of SD primarily in the PFC and thalamus, both of which are part of the attentional network involved in executive functioning. Similarly, the shift of encephalographic activity toward lower frequencies and the spatial shift of power toward frontal areas indicate, in all likelihood, reduced cortical arousal, particularly of the frontal areas.

Although neuroimaging and encephalography revealed mainly cortical effects of SD, residing predominantly in the PFC [104], little is known about the more detailed, microscopic processes that underlie those effects. When looking for potential sources, there is a longstanding agreement that the ascending nuclei in the reticular activating system are crucial for a normal sleep-wake cycle and are thus most likely involved in SD [cf. ref. 11]. The reticular activating system consists of several noradrenergic, dopaminergic, and cholinergic nuclei that project throughout the cortex and have a central role in attention and arousal [81, 114].In this review, we have mainly focused on the cholinergic system, as it is commonly believed that ACh is involved in attentional processes having SD-related effects on brain activity [cf. ref. 138]. Removal or blockage of cholinergic neurons reduces or disrupts the N1 and P300 components in encephalographic recordings. Because the cholinergic neurons innervate the thalamocortical network, which plays a central role in the generation of cortical rhythms, this cholinergic activity is likely related to the modification of cortical rhythms as observed during a sleep cycle and after SD. Several studies identified a strong increase of AD, particularly in the BF during SD, where AD has an inhibitory effect on the activity of the cholinergic nuclei [116118]. These results support the notion that AD is a modulator of the sleepiness associated with prolonged wakefulness as extracellular AD accumulates selectively in the BF and cortex and promotes the transition from wakefulness to slow-wave sleep by inhibiting cholinergic neurons in the BF [118].

Although this pathway is unlikely to be the only neural process affected by SD, it nicely illustrates how phenomena and processes at the distinguished scales (i.e., macro-, meso-, and micro-levels) might interact, and how dissection of those interactions might provide insight into the underlying causes of the decline in human functioning after SD. In this regard, it seems particularly interesting to investigate further the role of the PFC during SD, with the explicit aim to uncover the neural processes underlying its alleged compensatory function. After all, although there is ample evidence for the involvement of the PFC in executive functioning and SD on a macroscopic level, it remains unclear how these functions are instantiated at a microscopic level. For example, the PFC has top-down connections to the reticular activating system and would thus, in principle, be capable of modulating, or compensating for, the reduced activity in the cholinergic neurons.

On a final note, we would like to remark that the SD literature is replete with attentional studies, while studies of effects of SD on the performance of perceptual-motor tasks are scarce. In fact, our recent study on rhythmic force production [76] seems an exception in this regard. Perceptual-motor tasks are known to differ greatly in the extent to which they call upon attentional and higher-order executive processes, with great demands for entirely novel complex tasks and movement sequences and little or no demands for fully automated skills, such as maintaining balance. In this particular context, a secondary cognitive task may be readily introduced besides a primary perceptual-motor task to objectively assess the degree of automation of the latter task [158]. This may provide a proper benchmark for evaluating the neural effects of SD in terms of attention demands and attentive processes. We believe that such a strategy may open the way for investigating the putative roles of the attentional network in the planning and execution of automated and non-automated perceptual-motor tasks in an integrated fashion, i.e., across micro-, meso-, and macro-levels.