martes, 25 de enero de 2022

Consciousness: The last 50 years (and the next)

 

Anil K. Seth








 Brain and Neuroscience Advances

Volume 2: 1–6
© The Author(s) 2018
Article reuse guidelines: sagepub.com/journals-permissions DOI: 10.1177/2398212818816019 journals.sagepub.com/home/bna

Creative Commons Non Commercial CC BY-NC: This article is distributed under the terms of the Creative Commons Attribution-NonCommercial 4.0 License Http://www.creativecommons.org/licenses/by-nc/4.0/) which permits non-commercial use, reproduction and distribution of the work without further permission provided the original work is attributed as specified on the SAGE and Open Access pages (https://us.sagepub.com/en-us/nam/open-access-at-sage).

 

Introduction

The relationship between subjective conscious experience and its biophysical basis has always been a defining question for the mind and brain sciences. But, at various times since the beginnings of neuroscience as a discipline, the explicit study of consciousness has been either treated as fringe or excluded altogether. Looking back over the past 50 years, these extremes of attitude are well represented. Roger Sperry (1969), pioneer of split-brain operations and of what can now be called ‘consciousness science’ lamented in 1969 that ‘[m]ost behavioral scientists today, and brain researchers in particular, have little use for consciousness’ (p. 532). Presciently, in the same article he highlighted the need for new technologies able to record the ‘pattern dynamics of brain activity’ in elucidating the neural basis of consciousness. Indeed, modern neuroimaging methods have had a transformative impact on consciousness science, as they have on cognitive neuroscience generally.

Informally, consciousness science over the last 50 years can be divided into two epochs. From the mid-1960s until around 1990 the fringe view held sway, though with several notable exceptions. Then, from the late 1980s and early 1990s, first a trickle and more recently a deluge of research into the brain basis of consciousness, a transition catalysed by – among other things – the activities of certain high-profile scientists (e.g. the Nobel laureates Francis Crick and Gerald Edelman) and by the maturation of modern neuroimaging methods, as anticipated by Sperry.

Today, students of neuroscience – for the most part – feel able to declare (or deny) a primary interest in studying consciousness. There are academic societies and conferences going back more than 20 years and scholarly journals dedicated to the topic. Above all, there is growing a body of empirical and theoretical work drawing ever closer connections between the properties of subjective experience and the operations of the densely complex neural circuits, embodied in bodies embedded in environments that together give rise to the apparent miracle of consciousness. We cannot yet know whether today’s students will find the solution to the ‘problem of consciousness’ or whether the problem as currently set forth is simply misconceived. Either way, there is much more to be discovered about the relations between the brain and consciousness, and with these discoveries will come new clinical approaches in neurology and psychiatry, as well as a new appreciation of our place as part of, and not apart from, the rest of nature.

Consciousness: The last 50 years (and the next)

Anil K. Seth


Abstract

The mind and brain sciences began with consciousness as a central concern. But for much of the 20th century, ideological and methodological concerns relegated its empirical study to the margins. Since the 1990s, studying consciousness has regained a legitimacy and momentum befitting its status as the primary feature of our mental lives. Nowadays, consciousness science encompasses a rich interdisciplinary mixture drawing together philosophical, theoretical, computational, experimental, and clinical perspectives, with neuroscience its central discipline. Researchers have learned a great deal about the neural mechanisms underlying global states of consciousness, distinctions between conscious and unconscious perception, and self-consciousness. Further progress will depend on specifying closer explanatory mappings between (first-person subjective) phenomenological descriptions and (third-person objective) descriptions of (embodied and embedded) neuronal mechanisms. Such progress will help reframe our understanding of our place in nature and accelerate clinical approaches to a wide range of psychiatric and neurological disorders.

Keywords

Consciousness, neural correlates of consciousness, selfhood, volition, predictive coding
Received: 12 February 2018
Sackler Centre for Consciousness Science and School of Engineering and Informatics, University of Sussex, Brighton, UK

Corresponding author:

Anil K. Seth, Sackler Centre for Consciousness Science and School of Engineering and Informatics, University of Sussex, Brighton BN1 9QJ, UK.
Email: a.k.seth@sussex.ac.uk; www.anilseth.com
816019BNA0010.1177/2398212818816019Brain and Neuroscience Advances
Seth research-article2018

Article
Brain and Neuroscience Advances

Consciousness: 1960s until 1990

By the mid-1960s, the behaviourism that had dominated 20th century psychology (especially in America) was in retreat. A new cognitive science was emerging that recognised the existence and importance of inner mental states in mediating between stimulus and response. But consciousness as an explanatory target was still largely off-limits: ‘We should ban the word “consciousness” for a decade or two’ as cognitive scientist George Miller put it in 1962. As late as 1989, Stuart Sutherland (1989) wrote in the International Dictionary of Psychology: ‘Consciousness is a fascinating but elusive phenomenon. It is impossible to specify what it is, what it does, or why it evolved. Nothing worth reading has been written on it’. (p. 95).

Whatever its merits at the time, Sutherland’s judgement seems harsh in retrospect. Although lacking a coherent organisation or much coordination, the neuroscience of consciousness in this period saw several substantial advances. These advances, admittedly, were mainly about establishing correspondences between brain regions or activity patterns and properties of conscious experience, rather than addressing fundamental questions such as why consciousness is part of the universe in the first place. Good examples come from neurology and neuropsychology, where surgical interventions and the study of neurological illnesses, injuries and lesions were revealing the essential dependency of particular aspects of consciousness on specific brain properties.

Some of these early studies remain among the most thought provoking even today. The split-brain (more precisely: callosectomy) studies of Sperry and Michael Gazzaniga are a case in point. Their first experiments, on the World War II veteran W.J., revealed that each cerebral hemisphere could perceive a visual stimulus independently, with only the left hemisphere (in W.J.) being able to provide a verbal report (Gazzaniga et al., 1962). Follow-up studies found that the somatosensory system, the motor system, and many other perceptual and cognitive systems could be similarly ‘split’, while other systems – for example, emotion – remained intact (Gazzaniga, 2014). The extent to which a single cranium can house independent conscious subjects is still hotly debated (Gazzaniga, 2014; Pinto et al., 2017; Sasai et al., 2016) perhaps because the very idea challenges one of our most deeply held assumptions: that consciousness is necessarily unified.

The unity of consciousness was also challenged, more subtly, by psychosurgical lesions of the medial temporal lobe, which were carried out to alleviate intractable epilepsy. The case of Henry Moliason – patient H.M. – is well known to many. Following bilateral removal of the medial temporal lobe in 1953, including both hippocampi, H.M. was cured of his epilepsy but left with a profound anterograde (and substantial retrograde) amnesia (Scoville and Milner, 1957). In a series of studies, neuropsychologists Suzanne Corkin and Brenda Milner found that despite living in what Corkin (2013) called the ‘permanent present tense’, H.M. could learn new motor skills, had intact working and semantic memory, and was generally able to acquire a range of implicit (non-conscious) memories. Only his ability to acquire new explicit, conscious memories was affected. These findings not only charted a new topography of conscious and unconscious kinds of memory but they also showed how our apparently unified sense of conscious selfhood – in which episodic memories play a key role – can fragment so that some aspects persist while others are lost.

Another central feature of conscious selfhood is the experience of ‘free will’, or more precisely, experiences of volition (intentions to do this-or-that) and agency (being the cause of events). Here, the experiments of Benjamin Libet in the 1980s continue to inspire new research and to stoke controversy in equal measure. His studies, designed to measure the timing of conscious decisions to make voluntary movements, were based on a remarkably simple paradigm ((Libet, 1982) see Figure 1(a)). Participants press a button at a time of their own choosing and then report the time they felt the ‘urge’ to move – their conscious intention – by noting the position of a dot on an oscilloscope screen.

Libet first observed a previously described build-up of neural electrical activity prior to voluntary movement – the so-called ‘readiness potential’ ((Kornhuber and Deecke, 1965), (see Figure 1(b)). His key innovation was to show that this build-up started several hundred milliseconds before the participant was aware of their intention to move, challenging the assumption that the conscious ‘urge’ was the cause of the voluntary movement.

This interpretation has been debated ever since, sparking many fascinating experiments (Haggard et al., 2002; Schurger et al., 2012). Libet himself was uncomfortable with the idea that conscious intentions were epiphenomenal, suggesting instead that the time between the conscious ‘urge’, and the actual movement was sufficient to allow a conscious ‘veto’ to take effect. Any conscious ‘veto’, however, is also likely to have identifiable neural precursors – so this in itself does not resolve Libet’s metaphysical quandary (Brass and Haggard, 2007). Perhaps the most convincing interpretation of these puzzling phenomena is that experiences of intention and agency label particular actions – and their consequences – as being self-generated rather tan externally imposed, allowing the organism to learn and perhaps make better (voluntary) decisions in the future (Haggard, 2008).

 Consciousness: from the 1990s until the present

A convenient date to mark the rehabilitation of consciousness within neuroscience is with Francis Crick and Christof Koch’s (1990) landmark paper, ‘Towards a neurobiological theory of consciousness’, which opens with the line: ‘It is remarkable that most of the work in both cognitive science and the neurosciences makes no reference to consciousness (or “awareness”)’, and which goes on to propose a specific theory of visual consciousness based on gamma-band (~40 Hz) oscillations.

Although this specific idea has now fallen from favour, a new industry rapidly developed aimed at uncovering the so-called ‘neural correlates of consciousness’ (NCCs): ‘the minimal neuronal mechanisms jointly sufficient for any one conscious percept’ (Crick and Koch, 1990). The search for NCCs, boosted by the arrival of the now ubiquitous magnetic resonance imaging (MRI) scanner (as well as good-old-fashioned EEG, and invasive neurophysiology in non-human primate studies), gave consciousness research a pragmatic spin. Instead of worrying at the so-called ‘hard problem’ of how conscious experiences could ever arise from ‘mere’ matter (Chalmers, 1996), neuroscientists could get on with looking for brain regions or processes that reliably correlated with particular conscious experiences, or with being conscious at all.

Over the past quarter-century, there has been considerable progress in identifying candidate NCCs (Koch et al., 2016; Metzinger, 2000; Odegaard et al., 2017), at least in specific contexts like visual or auditory awareness. A classic method has been to compare brain activity for ‘conscious’ and ‘unconscious’ conditions, while keeping sensory stimulation (and as far as possible, everything else) constant. For example, in binocular rivalry, conscious perception alternates even though the sensory inputs (different images to each eye) remain the same. Early studies of the brain basis of binocular rivalry used implanted electrodes in monkeys trained to report which of two visual percepts was dominant.

These studies found that neuronal responses in early visual areas – in particular V1 – tracked the physical stimulus rather than the percept, while neuronal responses in ‘higher’ areas – like inferotemporal cortex (IT) – tracked the percept rather than the physical stimulus (Leopold and Logothetis, 1996; Logothetis and Schall, 1989). Subsequent human neuroimaging studies, however, found that neuronal activity in primary visual cortex did correlate with perceptual dominance (Polonsky et al., 2000), and the debate continues as to whether the neuronal mechanisms underlying perceptual transitions lie early in the visual stream, or in higher-order regions such as the parietal or frontal cortices (Blake et al., 2014). As well as rivalry experiments, so-called ‘masking’ paradigms have also been widely used in consciousness science. These paradigms enable comparison of supraliminal versus subliminal stimulus presentations in a variety of perceptual modalities, with many studies implicating activation of the fronto-parietal network in reportable conscious perception (Dehaene and Changeux, 2011; Figure 2). These two examples stand for many others; see Boly et al. (2017); Odegaard et al. (2017) for recent, and conflicting, reviews.

At the same time, a separate strand of research has focussed on transitions in conscious states, both reversible (e.g. sleep and anaesthesia (Massimini et al., 2005)) and following brain injury (e.g. coma and the vegetative state (Owen et al., 2009)). Here, the challenge is to identify the neural mechanisms that support being conscious at all, rather than those associated with being conscious of this or that. One difficulty here is that global transitions of this kind affect the brain and body very generally so that it is challenging to isolate the neural mechanisms underlying consciousness per se. There are additional difficulties in distinguishing so-called ‘enabling’ conditions from those neural mechanisms that actually support conscious states. For example, certain brainstem lesions can abolish consciousness forever, but many believe that the brainstem merely enables conscious states, while the actual ‘generators’ of consciousness may lie elsewhere (Dehaene and Changeux, 2011; though see Merker, 2007)).

New theories have accompanied these empirical developments. One of the most influential is Bernard Baars’ (1988) ‘global workspace’ theory, which proposes that modular and specialised processors compete for access to a ‘global workspace’. Mental states become conscious when they are ‘broadcast’ within the workspace so that they can influence other processes, including verbal report and motor action. More recent ‘neuronal’ versions of this theory associate the global workspace with highly interconnected fronto-parietal networks, linking conscious perception to a nonlinear ‘ignition’ of activity within these networks, a position in line with many neuroimaging studies (Dehaene and Changeux, 2011; Dehaene et al., 2003).

Workspace theories tend to interpret conscious perception in terms of ‘access’, in the sense that a percept is defined as conscious only if it is available for verbal (or other behavioural) report, as well as to other cognitive processes (memory, attention, and so on). An advantage of this view is that conscious status is readily assessable in experiments since conscious content is by definition reportable.

However, another common intuition is that perceptual or ‘phenomenal’ consciousness is ‘richer’ than we can report at any time, since reportability is limited by (among other things) the constraints of memory. The distinction between phenomenal consciousness and access consciousness (Block, 2005) remains a productive source of new experiments and controversy (Tsuchiya et al., 2015).

 The road ahead

These are exciting times in consciousness science, and there is only space here to gesture at some promising research directions. In terms of conscious level, new theories and measures have emerged based on ‘neuronal complexity’ and ‘integrated information’ (Seth et al., 2011; Tononi et al., 2016). The basic idea is that conscious scenes are both highly integrated (each is experienced as a unified whole) and highly informative (each conscious scene is one among a vast repertoire of alternative possibilities), motivating the development of mathematical metrics that combine the same properties. Excitingly, some practical approximations to these measures are showing promise in quantifying ‘residual’ awareness after brain injury, without relying on overt behaviour (Casali et al., 2013).

Brain and Neuroscience Advances

Research on conscious content has continued to focus on brain regions or processes that distinguish conscious from unconscious perceptions. Addressing a recurring concern about the NCC approach, new experimental paradigms are refining our understanding of interactions between the neuronal mechanisms underlying conscious perception, with those underlying behavioural report. So-called ‘no report’ paradigms, which infer perceptual transitions indirectly from (for example) automatic eye movements, are challenging the idea that frontal brain regions are constitutively involved in conscious perception ((Frassle et al., 2014), though see (Van Vugt et al., 2018) for evidence to the contrary).

At the same time, advances in analysis methods like signal detection theory (Green and Swets, 1966) are enabling researchers to draw more rigorous distinctions between objective and subjective aspects of perception, as well as to quantify individual differences in metacognition (cognition about cognition) that may be relevant to consciousness (Barrett et al., 2013; Fleming and Dolan, 2012).

Theoretically, ‘predictive coding’ or ‘Bayesian brain’ approaches stand to advance our understanding of the neural basis of conscious perception. These approaches model perception as a process of (possibly Bayesian) inference on the hidden causes of the ambiguous and noisy signals that impinge on our sensory surfaces (Friston, 2009). Inverting some classical views on perception, top-down signals are proposed to convey perceptual predictions, while bottom-up signals convey only or primarily ‘prediction errors’: the discrepancy between what the brain expects and what it gets at each level of processing. This framework provides a powerful interpretation of some relatively old findings associating conscious perception with the integrity of top-down signalling (Pascual-Leone and Walsh, 2001) and is motivating new studies which explicitly manipulate perceptual expectations, examining how this alters conscious perception (De Lange et al., 2018). Excitingly, these ideas may provide a mechanistic understanding of abnormal perception in some clinical contexts, a good example being the positive symptoms (e.g. hallucinations) of psychosis (Fletcher and Frith, 2009; Powers et al., 2017; Teufel et al., 2015).

There is also increasing focus in consciousness research on experiences of selfhood, which encompass basic experiences of embodiment and body ownership (Blanke et al., 2015), experiences of volition and agency (Haggard, 2008), as well as ‘higher’ aspects of selfhood such as episodic memory and social perception. Here, new developments in virtual and augmented reality (Lenggenhager et al., 2007; Seth, 2013), as well as in characterising interoception (the sense of the body ‘from within’ (Critchley et al., 2004)) are heralding new insights into how our apparently unified experience of being a ‘self’ is constructed, on the fly, from many potentially distinguishable sub-processes – and how breakdowns in this constructive process may underlie a variety of psychiatric conditions.

From this vantage point in time, 50 years after the birth of the British Neuroscience Association, it is fair to say that the scientific study of consciousness has regained its rightful place as a central theme in the mind and brain sciences. A great deal is now known about how embodied and embedded brains shape and give rise to various aspects of conscious level, content, and self. Of course, much more remains to be discovered.

Exciting new combinations of theory, experiment, and modelling are helping transform mere correlations into explanations that map from neural mechanism to phenomenology. Accompanying these developments are important clinical applications in neurology and psychiatry, as well as the Deep challenges of investigating consciousness in infancy, in other non-human animals, and perhaps even in future machines.

Whether or not the ‘hard’ problem of consciousness will yield to these and other developments, the history of the next 50 years will make for fascinating reading.

 

Declaration of conflicting interests

The author(s) declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.

 Funding

I am grateful to the Dr Mortimer and Theresa Sackler Foundation which supports the Sackler Centre for Consciousness Science. I also acknowledge support from the Wellcome Trust (Engagement Fellowship) and the Canadian Institute for Advanced Research, Azrieli Programme on Brain, Mind, and Consciousness.

References

Baars BJ (1988) A Cognitive Theory of Consciousness. New York: Cambridge University Press.

Barrett AB, Dienes Z and Seth AK (2013) Measures of metacognition on signal- detection theoretic models. Psychological Methods 18(4): 535–552.

Blake R, Brascamp J and Heeger DJ (2014) Can binocular rivalryreveal neural correlates of consciousness? Philosophical Transactionsof the Royal Society of London: Series B, Biological Sciences 369(1641): 20130211.

Blanke O, Slater M and Serino A (2015) Behavioral, neural, and computational principles of bodily self-consciousness. Neuron 88(1): 145–166.

Block N (2005) Two neural correlates of consciousness. Trends in Cognitive Sciences 9(2): 46–52.

Boly M, Massimini M, Tsuchiya N, et al. (2017) Are the neural correlates of consciousness in the front or in the back of the cerebral cortex? Clinical and neuroimaging evidence. Journal of Neuroscience 37(40): 9603–9613.

Brass M and Haggard P (2007) To do or not to do: The neural signature of self-control. Journal of Neuroscience 27(34): 9141–9145.

Casali AG, Gosseries O, Rosanova M, et al. (2013) A theoretically based index of consciousness independent of sensory processing and behavior. Science Translational Medicine 5(198): ra105.

Chalmers DJ (1996) The Conscious Mind: In Search of a Fundamental Theory. New York: Oxford University Press.

Corkin S (2013) Permanent Present Tense: The Unforgettable Life of the Amnesic Patient, H. M. London: Allen Lane.

Crick F and Koch C (1990) Towards a neurobiological theory of consciousness. Seminars in the Neurosciences 2: 263–275.

Critchley HD, Wiens S, Rotshtein P, et al. (2004) Neural systems supporting interoceptive awareness. Nature Neuroscience 7(2): 189–195.

De Lange FP, Heilbron M and Kok P (2018) How do expectations shape perception? Trends in Cognitive Sciences 22(9): 764–799.

Dehaene S and Changeux JP (2011) Experimental and theoretical approaches to conscious processing. Neuron 70(2): 200–227.

Dehaene S, Sergent C and Changeux JP (2003) A neuronal network model linking subjective reports and objective physiological data during conscious perception. Proceedings of the National Academy of Sciences of the United States of America 100(14): 8520–8525.

Fleming SM and Dolan RJ (2012) The neural basis of metacognitive ability. Philosophical Transactions of the Royal Society of London: Series B, Biological Sciences 367(1594): 1338–1349.

Fletcher PC and Frith CD (2009) Perceiving is believing: A Bayesian approach to explaining the positive symptoms of schizophrenia. Nature Reviews Neuroscience 10(1): 48–58.

Frassle S, Sommer J, Jansen A, et al. (2014) Binocular rivalry: Frontal activity relates to introspection and action but not to perception. Journal of Neuroscience 34(5): 1738–1747.

Friston KJ (2009) The free-energy principle: A rough guide to the brain? Trends in Cognitive Sciences 13(7): 293–301.

Gazzaniga MS (2014) The split-brain: Rooting consciousness in biology. Proceedings of the National Academy of Sciences of the United States of America 111(51): 18093–18094.

Gazzaniga MS, Bogen JE and Sperry RW (1962) Some functional effects of sectioning the cerebral commissures in man. Proceedings of the National Academy of Sciences of the United States of America 48(10): 1765–1769.

Green DM and Swets JA (1966) Signal Detection Theory. New York: Wiley.

Haggard P (2008) Human volition: Towards a neuroscience of will. Nature Reviews Neuroscience 9(12): 934–946.

Haggard P, Clark S and Kalogeras J (2002) Voluntary action and conscious awareness. Nature Neuroscience 5(4): 382–385.

Koch C, Massimini M, Boly M, et al. (2016) Neural correlates of consciousness: Progress and problems. Nature Reviews Neuroscience 17(5): 307–321.

Kornhuber HH and Deecke L (1965) Changes in the brain potential in voluntary movements and passive movements in man: Readiness potential and reafferent potentials. Pflügers Archiv für die gesamte. Physiologie des Menschen und der Tiere 284: 1–17.

Lenggenhager B, Tadi T, Metzinger T, et al. (2007) Video ergo sum: Manipulating bodily self-consciousness. Science 317(5841): 1096–1099.

Leopold DA and Logothetis NK (1996) Activity changes in early visual cortex reflect monkeys’ percepts during binocular rivalry. Nature 379(6565): 549–553.

Libet B (1982) Brain stimulation in the study of neuronal functions for conscious sensory experiences. Human Neurobiology 1(4): 235–242.

Logothetis NK and Schall JD (1989) Neuronal correlates of subjective visual perception. Science 245(4919): 761–763.

Massimini M, Ferrarelli F, Huber R, et al. (2005) Breakdown of cortical effective connectivity during sleep. Science 309(5744): 2228–2232.

Merker B (2007) Consciousness without a cerebral cortex: A challenge for neuroscience and medicine. Behavioral and Brain Sciences. 30(1): 63–81; discussion 81–134.

Metzinger T (ed.) (2000) Neural Correlates of Consciousness: Empirical and Conceptual Questions. Cambridge, MA: The MIT Press.

Odegaard B, Knight RT and Lau H (2017) Should a few null findings falsify prefrontal theories of conscious perception? Journal of Neuroscience. 37(40): 9593–9602.

Owen AM, Schiff ND and Laureys S (2009) A new era of coma and consciousness science. Progress in Brain Research 177: 399–411.

Pascual-Leone A and Walsh V (2001) Fast backprojections from the motion to the primary visual area necessary for visual awareness. Science 292(5516): 510–512.

Pinto Y, Neville DA, Otten M, et al. (2017) Split brain: Divided perception but undivided consciousness. Brain 140(5): 1231–1237.

Brain and Neuroscience Advances

Polonsky A, Blake R, Braun J, et al. (2000) Neuronal activity in human primary visual cortex correlates with perception during binocular rivalry. Nature Neuroscience 3(11): 1153–1159.

Powers AR, Mathys C and Corlett PR (2017) Pavlovian conditioninginduced hallucinations result from overweighting of perceptual priors. Science 357(6351): 596–600.

Sasai S, Boly M, Mensen A, et al. (2016) Functional split brain in a driving/listening paradigm. Proceedings of the National Academy of Sciences of the United States of America 113(50): 14444–14449.

Schurger A, Sitt JD and Dehaene S (2012) An accumulator model for spontaneous neural activity prior to self-initiated movement. Proceedings of the National Academy of Sciences of the United States of America 109(42): E2904–E2913.

Scoville WB and Milner B (1957) Loss of recent memory after bilateral hippocampal lesions. Journal of Neurology, Neurosurgery, and Psychiatry 20(1): 11–21.

Seth AK (2013) Interoceptive inference, emotion, and the embodied self. Trends in Cognitive Sciences 17(11): 565–573.

Seth AK, Barrett AB and Barnett L (2011) Causal density and integrated information as measures of conscious level. Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences: 369(1952): 3748–3767.

Sperry RW (1969) A modified concept of consciousness. Psychological Review 76(6): 532–536.

Sutherland S (1989) The International Dictionary of Psychology. New York, NY: Crossroads Classic.

Teufel C, Subramaniam N, Dobler V, et al. (2015) Shift toward prior knowledge confers a perceptual advantage in early psychosis and psychosis-prone healthy individuals. Proceedings of the National Academy of Sciences of the United States of America 112(43): 13401–13406.

Tononi G, Boly M, Massimini M, et al. (2016) Integrated information theory: From consciousness to its physical substrate. Nature Reviews. Neuroscience 17(7): 450–461.

Tsuchiya N, Wilke M, Frassle S, et al. (2015) No-report paradigms: Extracting the true neural correlates of consciousness. Trends in Cognitive Sciences 19(12): 757–770.

Van Vugt B, Dagnino B, Vartak D, et al. (2018) The threshold for conscious report: Signal loss and response bias in visual and frontal cortex. Science 360(6388): 537–542.


No hay comentarios.:

Publicar un comentario

33 grados a la sombra

Casi las 14:00 horas. Alrededor del mediodía desperté. Un silencio que no dice nada. En la lengua nicotina y cafeína. En cuerpo y pi...