Abstract
Within theoretical and empirical enquiries, many different meanings associated with consciousness have appeared, leaving the term itself quite vague. This makes formulating an abstract and unifying version of the concept of consciousness – the main aim of this article –into an urgent theoretical imperative. It is argued that consciousness, characterized as dually accessible (cognized from the inside and the outside), hierarchically referential (semantically ordered), bodily determined (embedded in the working structures of an organism or conscious system), and useful in action (pragmatically functional), is a graded rather than an all-or-none phenomenon. A gradational approach, however, despite its explanatory advantages, can lead to some counterintuitive consequences and theoretical problems. In most such conceptions consciousness is extended globally (attached to primitive organisms or artificial systems), but also locally (connected to certain lower-level neuronal and bodily processes). For example, according to information integration theory (as introduced recently by Tononi and Koch, 2014), even such simple artificial systems as photodiodes possess miniscule amounts of consciousness. The major challenge for this article, then, is to establish reasonable, empirically justified constraints on how extended the range of a graded consciousness could be. It is argued that conscious systems are limited globally by the ability to individuate information (where individuated information is understood as evolutionarily embedded, socially altered, and private), whereas local limitations should be determined on the basis of a hypothesis about the action-oriented nature of the processes that select states of consciousness. Using these constraints, an abstract concept of consciousness is arrived at, hopefully contributing to a more unified state of play within consciousness studies itself.
Keywords: taxonomy of consciousness, graded consciousness, extended consciousness, sensorimotor consciousness, integrated information, individuated information, subjectivity
“What do I call ‘information that it is raining’? (Or have I only information of this information too?) And what gives this ‘information’ the character of information about something?”
(Wittgenstein, 2009, I. §356)
Introduction
Over the last few decades, the research into consciousness carried on in both cognitive neuroscience and the philosophy of mind has led to the emergence and consolidation of dozens of divergent meanings for terms central to the topic (Brook and Raymont, 2006; Brook, 2008; Jonkisz, 2012). This polyvalence should surely count as surprising: after all, the resolving of basic conceptual issues would normally be expected to precede the experimental phase of an investigation. With consciousness studies, though, it would appear that questions about the meaning even of the central term itself get to be posed alongside those that pertain to its empirical status. Taking into account this issue alone, the formulation of an abstract concept of consciousness unifying the different branches of consciousness research – the overriding aim of this paper – would seem to represent an urgent theoretical imperative.
An initial step in the direction of meeting that challenge, proposed by the current author, was to assemble a stable framework of connections linking together the numerous varieties of consciousness currently encounterable in science: phenomenal and access, creature and state, sensorimotor and perceptual, reflexive consciousness and self-consciousness, minimal and normal, impaired and altered, clouded and epileptic, visual and auditory, bodily and social, animal and machine, etc. This was accomplished in the form of a proposed fourfold taxonomy (Jonkisz, 2012), by showing that virtually all such varieties can be understood as constituted with reference to just four aspects of the phenomenon: namely, its being dually accessible (both from the inside and from the outside, subjectively and objectively), hierarchically referential (semantically ordered), bodily determined (depending on physical states of an organism), and useful (possessing diverse pragmatic functions). Ultimately, the varieties of consciousness were accommodated within a taxonomy of epistemic kinds, semantic orders, physiological state categories, and the pragmatically differentiated types of consciousness supported (see Table 1).
Table 1.
The four-fold taxonomy of consciousness.
| Aspects of consciousness | Characteristic | Examples | |
|---|---|---|---|
| (Epistemic) Kinds of consciousness |
Subjective | Cognized from first-person perspective (experienced) | phenomenal, first-personal, qualitative, experiential |
| Objective | Cognized from third-person perspective (observed) | access, psychological, third-personal | |
| (Semantic) Orders of consciousness |
Sensorimotor | Refers to the environment. Enables basic, non-random motor responses utilized, e.g., in space navigation, maintaining equilibrium, etc. | Sensorimotor, ecological self, proto-self |
| Perceptual | Refers to percepts. Enables identification of the objects of perception and adjusting motor actions | perceptual, inner sense, first-order representation, primary | |
| Meta-perceptual | Refers to perception. Enables multimodal identification of the source of information and adjusting perceptual processes | HOT, HOP, reflexive, introspective… | |
| Self-consciousness | Refers to the perceiving subject. Enables identification of one’s own body and the ownership of the perceptual processes | self-consciousness, extended, reportable, secondary | |
| Meta-self-consciousness | Refers to one’s own self. Enables identification of an owner as an experiencing subject and formulation of the self-concept | symbolic self-concept, recursive self-consciousness | |
| (Physiological) States of consciousness |
Wakeful | Occur in physiologically normal wakefulness | NWS, creature, vigilance |
| Sleep-states | Occur in physiologically normal sleep | sleep, REM-consciousness, NREM… | |
| Impaired | Occur in neurological and psychological disorders (e.g., brain lesions) | LOC, minimal, blurred, clouded, delirious etc. | |
| Altered | Occur in ‘non-standard’ conditions (induced accidentally or intentionally) | ASC, hypnotic, trance, mindfulness meditation, etc. | |
| (Pragmatic) Types of consciousness |
Source-defined | Distinguished according to the type of receptor or sensory system (visual, olfactory,…) | visual, auditory, tactile, olfactory, gustatory, proprioceptive |
| Use-defined | Distinguished according to the type of situation it is used in | social, temporal, body, motor-skill, emotional, facial, etc. | |
| System-defined | Distinguished according to the type of subject in which it occurs | animal, human, artificial, machine. | |
*Revised version of table introduced in Jonkisz, 2012.
The current article argues that these four taxonomic aspects directly correspond to general or fundamental features of the phenomenon. It is also claimed that the broad picture arising from this taxonomy (which includes sensorimotor states within the scope of consciousness) implies a graded rather than an all-or-none nature for the phenomenon. Brief analysis of the empirical status of the graded versus all-or-none (or dichotomous) approaches prompts the conviction that the former is empirically justified, and preferable, when interpreting certain data and medical cases. However, the gradational approach may prove more problematic methodologically (e.g., contrastive analysis is more difficult to obtain; see Baars, 1994; Frith et al., 1999; Overgaard, 2004), and, more importantly, it may generate certain counterintuitive consequences and theoretical problems. Major instances of the latter arise from the fact that in most such theories graded consciousness is significantly extended globally (attached to significantly less developed organisms or even to artificial systems), but also locally (connected to certain lower-level neuronal and bodily processes). For example, according to a recently highly acclaimed theoretical model tailored to the gradational scenario, information integration theory (IIT; Tononi, 2004, 2008, 2010; Koch, 2012; Tononi and Koch, 2014), even such primitive artificial systems as photodiodes possess miniscule amounts of consciousness. Hence, one of the major theoretical challenges of this article is to establish reasonable limitations for how extended a graded form of consciousness could be.
The article concurs with the general assumption of IIT that all conscious states are informational states (differentiated and integrated), but above and beyond this the very notion of information will itself be analyzed from a biological perspective (since the phenomenon of consciousness has so far only been known to exist in vivo). As a result, information will be characterized as evolutionarily embedded, socially altered, and subjectively grounded – i.e., individuated. The notion of individuated information not only fosters a naturalized explanation of subjectivity (potentially rather significant in its own right), but also enables one to impose limitations on conscious systems (the global limitation postulate). Subsequently, it will be argued that since not all informational states accessible within a given organism or system are ultimately conscious (see Mudrik et al., 2014), certain additional selection mechanisms should be specified. Because there is no agreement over which of the cognitive functions, neuronal mechanisms, or brain areas are fully responsible for states of consciousness (e.g., attention, memory, language, or certain so called NCC’s, may each of them be neither sufficient nor necessary; see Noë and Thompson, 2004; Lamme, 2006, 2010; Hohwy, 2009), the article seeks to specify which of the general features of consciousness is most likely responsible for shaping the selection process. Consequently, one sort of limitation imposed on states of consciousness (the local limitation postulate) will be based on an empirically justified hypothesis about the pragmatically functional or action-oriented nature of the selection processes themselves. Finally, taking both local and global limitations as its basis, an abstract concept of consciousness will be arrived at. This conception enables one to link up together certain promising aspects of prominent neurocognitive models (e.g., IIT, GW, GNW) and philosophical ideas (e.g., aspects of ecological, embodied and enactive conceptions), and as such will hopefully contribute to a state of play within consciousness studies that will be (at least conceptually) more unified.
The initial stages of the present enquiry seek to give an adequate characterization of the fundamental features of consciousness. This is followed by analysis of the empirical status of the contrasting all-or-none (dichotomous) and gradational approaches to that phenomenon. Then come sections aimed at setting reasonable global and local limitations on how extended graded consciousness can be. This leads on in turn to the formulation of an abstract concept of consciousness presented in the last part of the article.
Fundamental Features of Consciousness
We come to know about consciousness in two different ways: by experiencing our own conscious states, and via something else – namely, observations of certain behaviors and neuronal processes presented by conscious agents. The difference between such subjective and objective forms of knowledge is essentially epistemic (hence the category of ‘epistemic kinds of consciousness’ in the abovementioned, previously proposed taxonomy; see Jonkisz, 2012, Table 1). Philosophers used to argue that only consciousness and the phenomena it consists of (i.e., thoughts, perceptions, feelings, emotions, desires, sensations, etc.) are subjectively experienced – hence the thought that asking ‘What-is-it-like to have it?’ in connection with anything else is absurd. In a similar sense, all or most conscious states are thought to be accompanied by experiential qualities or qualia (Nagel, 1974; Jackson, 1982; Dennett, 1988; Searle, 1992, 2000a,b; Block, 1995; Chalmers, 1995, 1996; Kriegel, 2006). As a result, the subjective (qualitative, experiential, phenomenal) nature of consciousness stands exposed as absolutely the hardest, and supposedly most enigmatic, aspect of the problem – one that engenders the so called explanatory gap faced by scientific approaches, and also perhaps even an ontologically unbridgeable gulf (Wittgenstein, 2009; McGinn, 1989; Chalmers, 1995, 1996; Levine, 1983, 2001; Bayne, 2009). Consequently, the distinctive feature of the phenomenon is often taken to be either our split access to it or the fact of its possessing an exclusively subjective character, where such considerations are also then viewed as furnishing a basis for antireductionist lines of argumentation and the drawing of dualistic consequences. (Such epistemico-ontological inferences have been present in the philosophy of mind at least since the time of Descartes’ distinction between res cogitans and res extensa; see Northoff and Musholt, 2006.) In this article, only the former strategy will be embraced, as the present author holds that dual epistemic accessibility does indeed help to distinguish consciousness from various other phenomena observed by science, and so should be regarded as indicative of one of its fundamental features.
Consciousness may be described as a relation between a given subject and some particular information or content he or she is aware of or simply possesses. In this sense, it is ‘directed toward’ something or simply ‘refers to’ or ‘is about’ something. It is with much the same considerations in mind that consciousness is also labeled either transitive (Rosenthal, 1986) or intentional (Searle, 1992, 2000a). Yet the latter notions mainly figure in philosophical debates, where ‘intentionality’ in particular, thanks to its ubiquitous deployment across phenomenology, analytical philosophy of perception and philosophy of mind, now seems too complex and ambiguous to be useful (see Pierre, 2003; Siewert, 2006; Lycan, 2014). In this context it is worth adding that consciousness may also refer – and quite often does – to its own content, in that it ‘reflects’ this in higher-order content, creating a hierarchy of orders (in the sense that a given subject X may not only be ‘conscious of Y,’ but also ‘aware of being conscious of Y’). The hierarchy is to a certain extent visible in so called higher-order theories of consciousness, or HOT’s (see Rosenthal, 1986; Gennaro, 2005), in various forms of self-directed perception (introspection or inner sense; see, e.g., Lycan, 1996), and in cases of metacognition (see Cohen and Dennett, 2011; Overgaard and Sandberg, 2012; Mealor and Deines, 2013; Shea et al., 2014). More abstract orders of consciousness are known to be correlated with higher levels of information processing in the brain, such as engage progressively more complex neuronal structures and enable the production of more complex forms of behavior: from basic sensorimotor reactions, perceptual differentiations and multimodal coordination to self-identification, verbal reports and the formulation of an abstract notion of selfhood (Hochstein and Ahissar, 2002; Crick and Koch, 2003; Baars, 2012; Palmer and Ramsey, 2012; Baars et al., 2013; Windey et al., 2013, 2014). Simplifying the idea somewhat, consecutive orders of consciousness may refer directly to the environment, to the objects perceived, to the perception itself, to the very perceiver, and/or to the self-conscious subject (see Table 1 and, for more detailed description, also Jonkisz, 2012). The semantic aspect of hierarchical referentiality counts as yet another fundamental feature of the phenomenon.
There exists an extensive body of scientific evidence directly correlating different states of consciousness – alongside their divergent impairments and various other alterations – with specific physiological and behavioral parameters as exhibited by some given conscious agent or other (e.g., Tart, 1969; Teasdale and Jennett, 1974; Bogen, 1995; Metzinger, 2000; Alvarez-Silva et al., 2006; Moller et al., 2006; Kokoszka, 2007; Schnakers, 2008; van Gaal et al., 2011; Långsjö et al., 2012; Panagiotaropoulos et al., 2012; Kouider et al., 2013). In this sense, the claim that all conscious states are determined by the interaction between ‘working’ structures or subsystems of a given organism’s body and changing environmental stimuli is quite natural. The fact that consciousness is bodily determined counts as another fundamental feature of the phenomenon. The chain of causal effects that are known to influence consciousness is traced back and forth, from the bottom–up molecular and neuronal activations, to the general neurophysiology and, further, to any biological processes or behavioral activities presented by a given conscious body, or specific environmental conditions it is faced with. Limiting ourselves to the biological body proves problematic when taking into account the possibility of non-biological conscious systems like those figuring in the debate over artificial or machine consciousness (Hollande, 2003; Torrance et al., 2007; Kurzweil, 2012; O’Regan, 2012). It is rather implausible to call states presented by those systems biological or physiological, yet their consciousness, if they possess it, should share the same fundamental features. In the light of this, it would be less question-begging to say that states of consciousness depend on a variety of physical or structural changes that go on within a conscious system as it interacts with its environment. On the other hand, however, the four state categories featured in the taxonomy (i.e., wakeful, sleep, impaired, and altered states of consciousness, see Table 1, Jonkisz, 2012) relate directly to the phenomenon of consciousness as it is currently studied in humans or animals, so some reference to the physiology (as a causal intermediary in the larger causal chain) would seem natural.
It seems that consciousness must have been an advantageous trait where the adaptive strategies of our ancestors were concerned (see, for example, Feinberg and Mallatt, 2013). From this perspective, the evolutionary advantageousness of the ability to be conscious may be characterized as usefulness in action, in the sense that consciousness fulfills certain pragmatic functions – e.g., enabling one to react appropriately, make specific choices, solve new problems, change one’s mind, or plan actions in advance, anticipate developments, etc. It is worth noting here that the notion of pragmatic function has much in common with Millikan’s (2004) idea of a proper function, as the latter may also be characterized as a form of biologically justified utility common to the creature and its ancestors. One might say that the functions of consciousness are structured and limited by at least three factors: the source of information (type of receptor, sensory, and memory systems), the conditions of application (whether the conscious information is applied to social life, spatial movement planning, or to temporal segregation of events, etc.), and the type of creature or system considered conscious (whether it is human, animal, or machine) – hence the taxonomic division into source, use, and system-defined types of consciousness (see Table 1, Jonkisz, 2012). On top of these, usefulness in action, defined in terms of pragmatic functionality, can then be considered a fourth fundamental feature of the phenomenon of consciousness.
Of course the facts, features, and characteristics ascribed to consciousness are highly varied: for example, that it feels a certain way, is phenomenally structured, is integrated, is a higher-order or self-representational state, is generated by recurrent processing in the thalamocortical structures, is associated with reverberating activations of the parieto-frontal areas, that it enables one to behave appropriately, adapt to new situations, learn, understand emotions, differentiate, and choose, etc. (Rosenthal, 1986; Baars, 1994; Block, 1995; Chalmers, 1995; Lycan, 1996; Damasio, 1999; Edelman, 2003; Tononi, 2004; Gennaro, 2005; Kriegel, 2006, 2007; Lamme, 2006; Dehaene and Changeux, 2011, etc.). It is hoped, however, that all of those just mentioned, as well as most other characteristics, ultimately fall into one of the four categories described above, as they concern facts that will prove to be either epistemic, semantic, bodily physiological, or functional. The four fundamental features that enable us to describe consciousness as being dually accessible, (hierarchically) referential, bodily determined, and useful in action may therefore be seen as together furnishing an abstract meta-characterization of the phenomenon. Somewhat more precisely, we might say that according to this view, consciousness is an objectively observable and subjectively experienced phenomenon that enables a given agent to utilize increasingly abstract contents. The increasing complexity of the contents and functions of consciousness (depicted in the taxonomic hierarchy of orders – see Table 1) is known to be correlated with the higher levels of information processing in the brain that engage progressively more complex, more wide-ranging neuronal structures and significantly more synchronized connection patterns (Hochstein and Ahissar, 2002; Crick and Koch, 2003; Baars, 2012; Baars et al., 2013; Windey et al., 2013, 2014). From this perspective, consciousness shows up as a graded rather than all-or-none phenomenon (with such gradedness being visible both in the referential contents of consciousness and in the underlying neuronal processes). If certain sensorimotor processes are to be included within its scope (as is assumed in the taxonomy), then the phenomenon also shows up as vastly extended among living creatures (since non-random sensorimotor activity is exhibited even by very simple organisms). Just how great the range of organisms is across which consciousness may be said to extend is something that is very much debated in the scientific literature, with few wishing to limit subjective consciousness to humans (Carruthers, 1998) and some ascribing it to quite a variety of non-human animals (Griffin, 2001; Griffin and Speck, 2004; Seth et al., 2005; Edelman and Seth, 2009; Boly et al., 2013; Feinberg and Mallatt, 2013), or even to artificial systems (Hollande, 2003; Torrance et al., 2007; Kurzweil, 2012; O’Regan, 2012). Both the graded character and the range of extension of consciousness are issues analyzed further in subsequent sections here.
Dichotomous and Graded Consciousness
In experimental practice a given subject is usually only deemed conscious of certain sensory stimuli, perceptual contents, emotions, thoughts, etc., when they (or it) can ‘prove it,’ either through relevant and statistically significant behavioral responses or in the form of conclusive reporting of their own states. In other words, consciousness is equated with either objectively or subjectively measurable forms of awareness. The current debate in science over the adequacy of objective versus subjective measures of consciousness (Seth et al., 2008; Overgaard et al., 2010; Sandberg et al., 2010; Szczepanowski et al., 2013; Wierzchoñ et al., 2014) to some extent mimics philosophical arguments over access versus phenomenal consciousness. Thus, if phenomenal consciousness (accompanied by subjective qualia) is to indeed be the main and most difficult target of explanation, as some philosophers insist (e.g., Nagel, 1974; Jackson, 1982; McGinn, 1989; Block, 1995; Chalmers, 1995, 1996; Levine, 2001; Kriegel, 2006; Bayne, 2009), then subjective measures would seem to be fundamentally more adequate (Sandberg et al., 2010, 2011; Szczepanowski et al., 2013; Peremen and Lamy, 2014). However, as the methodologies for subjective measuring assume reportable states to be decisive, consciousness may seem to be ‘switching on’ from unconsciousness rather late in the process – i.e., not until a given subject is actually able to present such metacognitive reports. Although metacognition, unlike verbal reportability, is not an exclusively human-related ability (Smith, 2009), it is associated with sophisticated stages of information processing, which makes the picture of consciousness appear to be almost exclusively human-related and so rather rare in nature. Often consciousness is viewed not just as coming into play quite late on, but also as doing so suddenly – as an all-or-none dichotomous neuronal and phenomenal event (Sergent and Dehaene, 2004; Dehaene et al., 2006). The following sections aim to show that the opposite view, according to which consciousness is a non-dichotomous, graded phenomenon, and rather widespread in the animal world, is not only well justified empirically but also explanatorily advantageous (see also Overgaard et al., 2006).
The view of consciousness as dichotomous has garnered substantial support across the field (e.g., Merikle and Reingold, 1998; Merikle et al., 2001; Sergent and Dehaene, 2004; Dehaene et al., 2006; van Gaal and Lamme, 2012; Balconi and Bortolotti, 2013; Landry et al., 2014). There have also been experimental attempts at showing that conscious and unconscious processes are biologically ‘bifurcated’ in that they rely on different neurological mechanisms (Sergent and Dehaene, 2004; Brogaard, 2011; van Gaal et al., 2011; Panagiotaropoulos et al., 2012). It must also be acknowledged that within the currently prevalent methodology of contrastive analysis (Baars, 1994; Frith et al., 1999), it is much easier and more effective in practice to demonstrate major differences in neural functioning between higher-order metacognitive states and certain lower-order non-reportable states such as would be considered unconscious, than to detect continuous, small changes, or to show ‘in-between’ states (Overgaard et al., 2006). Even so, the all-or-none picture of consciousness itself is more likely to run into considerable difficulties when it comes to interpreting experimental and medical data. The following cases may be briefly invoked as furnishing pertinent examples: interpretations of Libet-type experiments as examples of strictly unconscious, non-voluntary decision making (Libet, 1985; Soon et al., 2008); cases where the blindsight phenomenon is treated as genuinely unconscious perception (Ro et al., 2004; Brogaard, 2011); situations where non-human creatures unable to produce higher-order states are said to be totally bereft of consciousness (Carruthers, 1998); assessments of divergent impaired states of consciousness in the dichotomous scenario – according to Giacino (2005), for example, over just 12 years up to 41% of patients inhabiting a minimally conscious state were misdiagnosed (see Schnakers, 2008). In all these cases it is debatable whether we are dealing with an absence of consciousness in its entirety or just of certain aspects or higher forms of it (e.g., the ability to report, to discriminate objects, to recall and react accordingly; see below). Positing a dichotomous split between conscious and unconscious processes may also entail certain debatable metaphysical consequences. For example, if a given state does not count as conscious at all until it has become the object of some higher-order or metacognitive state, consciousness may start to look like an extra added property – the sudden by-product of unconscious brain processes, from which emergentistic or even epiphenomenalistic conclusions might then have to be drawn. So called conceivability arguments involving hypothetical zombie-creatures (observably identical to us but totally lacking consciousness – see Kirk, 1974, 2005; Chalmers, 1996, 2002; Stoljar, 2007) are clearly much more readily formulated in such a context. With the all-or-none picture in mind, consciousness is also more likely to be seen as a certain thing or entity that is either possessed or not (as life was seen by vitalists) – something which may eventually prompt one to embrace even more radical dualistic consequences regarding its nature. (A more detailed discussion of the metaphysical consequences of the all-or-none approach unfortunately lies well beyond the scope of this article.)
Non-dichotomous approaches, such as portray consciousness as a hierarchical or a graded process, are at least equally popular in the current context of debate (Natsoulas, 1983, 1997a,b; Neisser, 1988; Cleeremans, 1994, 2011; Damasio, 1999; Cleeremans and Jiménez, 2002; Ramsøy and Overgaard, 2004; Tononi, 2004, 2008, 2010; Morin, 2006; Overgaard et al., 2006, 2010; Seth et al., 2008; Kouider et al., 2010; Nieuwenhuis and de Kleijn, 2011; Koch, 2012; Feinberg and Mallatt, 2013; Miller and Schwarz, 2014; Peremen and Lamy, 2014). What is more important, though, is the high level of plausibility those conceptions have, in the light of both considerable supporting evidence and of their explanatory advantages in respect of the interpretation of experimental and medical data – both points to be briefly demonstrated below.
Miller and Schwarz (2014) review the number of behavioral and neurophysiological studies providing extensive support for the gradational view of consciousness, which the authors then apply to the analysis of the decision making process. They conclude that decisions are more likely to be reached in a gradual process of evidence-accumulation involving ascending certainty-levels (see also Mele, 2010), and a gradually acquired awareness of their being made, than they are to happen suddenly, as products of totally unconscious brain activity with delayed, epiphenomenal consciousness only being acquired later on in the process (Libet, 1985). The results fits well with currently popular methodologies, such as the perceptual awareness scale (PAS) proposed in (Ramsøy and Overgaard, 2004), the experimentally supported continuum of clarity levels described in (Overgaard et al., 2006, 2010), post-decision wagering (Nieuwenhuis and de Kleijn, 2011), and other similar strategies (Massoni et al., 2014; Wierzchoñ et al., 2014).
Further evidence and arguments in support of a hierarchical view of consciousness may be found in neurology, neuroanatomy and psychiatry, as well as in evolutionary and developmental approaches. Gradedness is clearly visible in the way the neuronal processes associated with consciousness unfold. According to the most common view, strong reciprocal connectivity between thalamic nuclei (especially the intralaminar and medial ones) and the relevant cortical areas enhance gradually more synchronized, more frequent and longer lasting activities in circuits connecting more distant groups of neurons (see, for example, Edelman and Tononi, 2000; Lamme, 2006; Edelman et al., 2011; Baars, 2012; Baars et al., 2013; Saalmann, 2014). The experiments described in Fleming et al. (2010) show significantly higher introspective accuracy in ratings of one’s own conscious decisions in subjects whose brains possess a greater volume of gray matter in the anterior prefrontal cortex region. Such sophisticated introspective abilities, realized in the evolutionarily youngest parts of the brain, may be viewed as representing the most advanced grades of self-awareness, and also as revealing their phylogenetic immaturity and phenotypic divergences in the human brain. Gradedness is also visible in the descriptions of different states of impairment. In medical practice the range of impairments and recovery prognoses are assessed against specified scales: e.g., the Glasgow Coma Scale and its revisited version (GCS, GCS-R), the Coma Recovery Scale and Revisited (CRS, CRS-R), the Full Outline of UnResponsiveness scale (FOUR), or the Wessex Head Injury Matrix (WHIM; see Giacino, 2005). Patients’ states are analyzed with respect to arousal level and behavioral responsiveness, creating a list of successively diminished levels of consciousness with minimal consciousness at the end of the list, just before vegetative state and coma (Teasdale and Jennett, 1974; Giacino, 2005; Schnakers, 2008). Other studies where a non-dichotomous hierarchical approach would seem particularly promising show, for instance, the following: that the return from anesthesia also occurs in steps, with so called primitive consciousness emerging first (Långsjö et al., 2012); that signs of diminished consciousness are revealed even in severely damaged brains (Philippi et al., 2012); that the cerebral cortex is probably not necessary for certain basic forms of consciousness (Merker, 2007); that consciousness occurs, to a certain degree, at different developmental stages (Stuss and Anderson, 2004; Zelazo, 2004); that even very young infants might be perceptually conscious (Kouider et al., 2013); that consciousness is not only a feature of the human mind, as animal consciousness most likely extends beyond the mammalian kingdom (Lyvers, 1999; Griffin, 2001; Reiss and Marino, 2001; Griffin and Speck, 2004; Seth et al., 2005; Legrand, 2007; Edelman and Seth, 2009; Smith, 2009; Boly et al., 2013; Feinberg and Mallatt, 2013).
Information integration theory, as developed recently by Tononi (2004, 2008, 2010), Koch (2012), Tononi and Koch (2014) is an abstract, theoretical model tailored to the gradational scenario: “The IIT claims that consciousness is not an all-or-none property, but is graded: specifically, it increases in proportion to a system’s repertoire of discriminable states” (Tononi, 2008, p. 236). Roughly speaking, according to IIT, a given system is more conscious the more chunks of information it is able to differentiate and bind together or integrate, where “Integrated information (Φ) is defined as the amount of information generated by a complex of elements, above and beyond the information generated by its parts…” (Tononi, 2008, p. 216). Certain problematic consequences and promising aspects of this model will be discussed in the sections below.
Extended Consciousness and its Limitations
On the basis of the evidence reviewed above, the hypothesis that consciousness is acquired in a graded manner rather than all at once definitely appears valid. If this is the case, it is also likely that at least some of the processes traditionally deemed unconscious rely on the same neurobiological mechanisms as subsequent conscious processes (Peremen and Lamy, 2014), being just an antecedent, lower-level, lower-order, degraded (etc.) form of these. Such an extension of consciousness into processes hitherto considered unconscious receives support, for example, from Overgaard et al. (2008), where the blindsight phenomenon is described as a degraded consciousness or, even more directly, from Holender and Duscherer (2004, p. 872), where a radical paradigm shift is proclaimed by the statement that “…dual-process models, which are based on both conscious and unconscious perception, should be rejected in favor of the single-process conscious perception model.”
Generally, it may be said that in gradational approaches consciousness is most often extended below the level of reportable and metacognitive states, to the point where not only lower-order perceptual processes, but also, sometimes, certain non-perceptual processes are reckoned to lie within its scope. Indeed, the latter is the case with this author’s own taxonomic hierarchy of orders (Jonkisz, 2012), in which consciousness is extended beyond perception toward sensorimotor processing. For example, when a given creature effectively avoids obstacles by utilizing visual, tactile, auditory or any other sensory information available in the environment, such functioning might be regarded as an early manifestation of consciousness. In performing such actions with above chance efficiency levels, the creature must be in possession of at least rudimentary sensorimotor information about the object’s surface and location, and about its own body position: on this basis alone it could be said that the creature possesses first-order consciousness of the environment and of its own body, and this even if its conscious states are not available for any higher-order monitoring processes (Kriegel, 2007; Jonkisz, 2012). The attribution of consciousness to certain lower-order states may be called local extension. By analogy, global extension would be an appropriate label to apply when consciousness is being attached to significantly less developed species than humans or primates (Griffin, 2001; Griffin and Speck, 2004; Seth et al., 2005; Edelman and Seth, 2009; Boly et al., 2013; Feinberg and Mallatt, 2013).
In dichotomous approaches, not only are certain non-reportable perceptual states regarded as unconscious (Kihlstrom, 1987; Sergent and Dehaene, 2004; Dienes and Scott, 2005; Dehaene et al., 2006; Scott and Dienes, 2008, 2010; Panagiotaropoulos et al., 2012; Landry et al., 2014), but also certain higher-order processes sometimes get to be placed outside of conscious awareness (van Gaal and Lamme, 2012). Having excluded a great many states, it seems easier for such conceptions to state what consciousness is not than to specify exactly what it is. Meanwhile, in the non-dichotomous gradational views, such as include a whole range of states into the conscious realm, the opposite concern may emerge as equally problematic: namely, the issues of how to determine which states are definitely not conscious. Consequently, no matter which approach is at stake, the methodology of contrastive analysis (Baars, 1994; Frith et al., 1999) seems destined to suffer permanently within the science of consciousness, as clear-cut contrasts and precise conditions or strict definitions are still rather difficult to obtain there (Overgaard, 2004). In setting a definite metacognitive or neuronal threshold for consciousness (represented, for example, by reportable states or recurrent processing), dichotomous approaches not only struggle to prove that this is the crucial threshold, but also face the burden of providing us with an unambiguous non-operational characterization of what, exactly, it is that is added when the threshold finally gets crossed and consciousness “switches on.” Such a definite borderline for consciousness should be considered dubious on both neuronal and phenomenal grounds: as was already mentioned, it is quite likely that certain perceptual processes traditionally considered unconscious in fact rely on the same neurological mechanisms as those associated with consciousness (Peremen and Lamy, 2014), while there is also strong experimental evidence for gradedness in subjective ratings of conscious contents and decisions with respect to awareness, clarity, certainty, or confidence levels (Overgaard et al., 2006, 2008, 2010; Mele, 2010; Nieuwenhuis and de Kleijn, 2011; Szczepanowski et al., 2013; Miller and Schwarz, 2014; Massoni et al., 2014; Wierzchoñ et al., 2014). In seeking to extend consciousness both locally and globally gradational conceptions struggle, on the other hand, to give unequivocal answers to a number of questions. Which stages of information processing (metacognitive, perceptual, or sensorimotor) form the final frontiers of consciousness? What types of neuronal activity (e.g., recurrent or feed-forward activations) and which areas of the brain (e.g., only the corticothalamic complex, or also certain limbic and midbrain regions) form the most basic correlates of consciousness? Does consciousness occur only in the mammalian kingdom or across all animal species? What about even more primitive living systems, like plants or bacteria? Could they possess any of its forms? And what about artificial systems (Hollande, 2003; Torrance et al., 2007; Kurzweil, 2012; O’Regan, 2012)? Could any sort of information be accompanied to some degree by consciousness (Chalmers, 1995; Tononi, 2008; Koch, 2012)?
Is it at all feasible to try to impose reasonable, empirically validated constraints on how far graded consciousness might extend? Or could it be that the only option that remains when adopting the gradational approach is a radical extension according to which almost all the information-bearing states of all information-processing systems, either biological or artificial, would be conscious to certain extent? The double-aspect principle introduced by Chalmers (1995, 1996), and the IIT proposed by Tononi (2004, 2008, 2010) and Koch (2012), have pursued just this radical strategy. Chalmers (1996) claims that “…wherever there is a causal interaction, there is information, and wherever there is information, there is experience” [by ‘experience’ he means phenomenal consciousness]. He also believes that “Experience is information from the inside; physics information from the outside” (Chalmers, 1996, p. 297 and 305). According to Koch (2012, p. 131)“[a]ny system at all that has both differentiated and integrated states of information…” is more or less conscious. As is commonly known, these theses result in strikingly counterintuitive consequences, such as the attribution of consciousness even to such simple artifacts as thermostats or photodiodes: “Strictly speaking, then, the IIT implies that even a binary photodiode is not completely unconscious, but rather enjoys exactly 1 bit of consciousness” (Tononi, 2008, p. 236). Also rather unpopular are the metaphysical consequences of the claims: i.e., certain forms of dualism or panpsychism that lead to a reorganization of the ontological structure of the universe with conscious information recognized as another fundamental entity besides mass or energy (Chalmers and Searle, 1997; Koch and Tononi, 2013; Searle, 1997, 2013a,b).
The authors of IIT have acknowledged the difficulties involved, and have moderated certain aspects of their theory. In particular, they propose a “set of postulates concerning requirements that must be satisfied by physical systems to account for experience” (Tononi and Koch, 2014, p. 4), but Searle’s (2013b) common sense critique of the theory remains sound, at least in some cases. At the same time, Searle might well be wrong where the very notion of information is concerned, this being for him something that cannot be used “to explain consciousness because the information in question is only information relative to a consciousness.” Searle (2013b) seems to be operating with a rather naive or colloquial conception of information (“…information is carried by a conscious experience of some agent (my thought that Obama is president, for example)…;” whereas Koch and Tononi (2013) try to adapt the formal, quantitative conception to the qualitative and unified character of conscious experience. In IIT, information may be approximately described as ‘any difference discernible from a perspective intrinsic to a given system’: “Consciousness… is both integrated (each experience is unified) and informative (each experience is what it is by differing, in its particular way, from trillions of other experiences). IIT introduces a novel, non-Shannonian notion of information – integrated information – which can be measured as ‘differences that make a difference’ to a system from its intrinsic perspective, not relative to an observer” (Koch and Tononi, 2013). If we acknowledge the obvious claim that consciousness consists of states that a given system is able to differentiate from its intrinsic perspective, then we must also agree that consciousness always carries certain information (in the sense of IIT). Consequently, if all conscious states are informational states, then IIT’s notion of information would indeed appear like a good candidate for the general characteristic of consciousness, contra Searle. On the other hand, however, the fact that all conscious states bear information does not itself entail that all informational states are conscious – contra Tononi (2008, p. 236) and Chalmers (1996, p. 297), for whom information or integrated information is coextensive with consciousness. Such assumptions “strongly violate our intuition…,” as they lead directly to the attribution of consciousness to almost any system possessing informational states – and thus support, in particular, the claim that a thermostat or a photodiode “…feels like something…” (Tononi and Koch, 2014, p. 7).
Information integration theory’s notion of integrated information, specifically understood as a simultaneously differentiated and unified state for a given system, fits our intuitions about consciousness, as it seems, indeed, to be composed of such states. However, even if the claim that all conscious states carry some information for the subject is justified and true, it does not necessarily hold the other way around – contrary to IIT, we claim that the question of whether all integrated information is conscious remains to a certain degree open. If we seek to claim otherwise, as IIT itself does (Tononi, 2008), then the characterization of consciousness may be threatened with circularity: ‘consciousness’ is defined in terms of ‘integrated information,’ ‘integrated information’ in terms of ‘differentiated and unified states,’ and ‘differentiated and unified states’ in terms of being conscious. In fact, we may still ask whether it is possible for a given state to be differentiated by a given system, and perhaps integrated, but still not conscious. (That is, we may ask whether it is empirically rather than logically possible; language itself will not, after all, tell us anything new about the real phenomena involved; see Chalmers, 2002). Koch himself (in Mudrik et al., 2014) has recently softened his stance and reconciled himself to the claim that this is indeed the case, at least in four cases: when the spatiotemporal window for processing is too narrow, when the level of semantic complexity of the information is low, when sensory information need not be detected multimodally, and/or when certain stimuli or states are already well known and so not novel for the subject. If all conscious states are integrated informational states, but not all states of this sort are conscious, then information integration is not sufficient for consciousness, although most likely it is necessary. Therefore a theory construed in informational terms will have to invoke additional selection mechanisms, or spell out in some other way the special conditions sufficing for consciousness to be present. Hopefully, such a theory will prove feasible and potentially consistent with major empirical findings without obliging one to relinquish the gradational scenario.
The aim of the following sections is to place limitations on what graded consciousness could amount to, thereby avoiding the improbable consequences of IIT’s commitment to radical extension, while nevertheless keeping the notion of ‘information’ in play – as one that remains promising where any general characterization of consciousness is concerned.
Individuated Information and Global Limitations
If all conscious states are indeed informational states (as was argued above), then it could be worth looking more closely at the nature of information itself. In the context of consciousness studies, it seems reasonable to characterize information from a naturalistic, biological perspective, since up to now the phenomenon of consciousness has only been known to exist in vivo.
Biological systems evolve in interaction with their environment and with other creatures; as a result, their bodies and bodily structures are attuned or sensitive to the specific informational resources available in the ecological niche of their ancestors (e.g., certain wavelengths of light, acoustic frequencies, or chemicals (etc.) detected on the ground, in the air, underwater, and so on). In this sense, it may be said that the information accessible to a given creature is evolutionarily embedded or phylogenetically determined – and, as far as we know, it is also epigenetically modified across different phenotypes (Migicovsky and Kovalchuk, 2011). The ways in which available information is interpreted by living creatures may also be socially altered, in the sense of being slightly modified across different groups of individuals even within the same species: e.g., variations in mating habits, ways of communicating danger, ways of hunting, interacting in a group, speaking, etc. (Swaddle et al., 2005). Moreover, the information actually utilized by living creatures is always subjectively grounded or private, in the sense of being determined by divergent subject-related factors. It is encoded in unique and dynamically changing patterns of neuronal activity, such as are modified by the actual physiological state and/or specific behavioral habits formed during individual histories (ontogenesis) and shaped by current environmental or social engagements, as well as by the particularities of the relevant location in space. The unique and private form of the informational state a creature experiences will reflect all these specific internal and external conditions – both those currently affecting the biological system at a given moment of time and those that did so in the past.
Such a view partakes of certain important ideas endorsed by ecological psychology. According to both Berker (1968) and Gibson (1979), cognition is always situated: i.e., it belongs to a specific ecological niche and depends on ecological information such as determines the set of possible actions (affordances). Perception, in this theoretical approach, is “based on [a given organism’s] detecting information” (Gibson, 1979, p. 1). The biologically focused stance presented here also has a lot in common with the conceptions (Millikan, 1989, 2004, 2006) has introduced. Her notions of ‘biological utility’ and ‘local natural information’ also stress the importance of the usability or applicability of the informational content determined by environmental conditions, by evolution and by individual histories. “The kind of knowledge that earthly creatures have is knowledge applicable in the domains they inhabit, not knowledge for arbitrary nomically possible worlds, nor for other domains, regions or eras within the actual world” (Millikan, 2004, p. 9). Embodied and enactivist theories also point to and stress similar aspects and determinants of the mind and of consciousness: “consciousness is an achievement of the whole animal in its environmental context” (Noë, 2010, p. 9). In this context, see also (Varela et al., 1991; Clark, 1997, 2008; Hurley, 1998, 2001; O’Regan and Noë, 2001; Thompson and Varela, 2001; Noë, 2006; O’Regan et al., 2006; Engel et al., 2013).
Within such accounts we see information being characterized in the way that has been described here by this author as evolutionarily embedded, socially altered, and subjectively grounded (private). Consequently, the ways in which living systems extract, store or interpret information are determined by their evolutionary antecedents, by interactions with other creatures, and by unique individual histories and actual situations. It may be concluded, therefore, that the form of information functionally relevant for a given biological system is always individuated – with respect to just that very system. Yet the philosophical implications of the latter statement are surely fairly momentous, given the epistemological – or possibly even ontological – consequences that may issue from it. In the former case, it may be claimed that individuated information is accessible only from-the-inside, from that system’s own point of view, so that it is epistemologically subjective. If one pursues the more far-reaching ontological implications as well, it may be insisted that the information only exists inside the system itself, making it ontologically subjective. (This is in fact the main point of contention between Searle and Dennett; see Dennett, 1991, 1993, 1995; Searle, 1992, 1995.) There is no need to address that metaphysical issue here, as for current purposes what is crucial is just the very notion of individuated information. According to the conception set forth, information functionally relevant for a given biological system is formed in the unique and particular interactions occurring between the system (e.g., its sensory organs located on its body and in relevant areas of its brain) and its environment (e.g., photons, sound waves, chemicals), or just between parts of the system itself, taken in isolation (e.g., its memory subsystems, insofar as these are able to induce certain states internally). Every aspect of the whole formation process possesses its own subject-related uniqueness: from the mere ability to move in certain ways, the sensitivity of its sensory organs, the anatomical and physiological characteristics of certain sensory areas, to its distinct location in some quite specifically ordered spatial milieu at a given moment of time and its individual engagements within a given ecological niche. So we may declare individuated information to be subject-related – or, simply, subjective (be it epistemologically or ontologically so).
The evolutionary mechanisms inherited in the context of our terrestrial ecology, and the complexity of the possible interactions between individual entities within it, have resulted in an immense diversity of unique biological systems. As far as we know, every organism is indeed more or less different. For example, even monozygotic twins reared together are actually far from identical: developmental changes are known to occur between them at virtually every possible level, be it genetic, epigenetic, anatomical, behavioral, etc. (Pfefferbaum et al., 2004; Fraga, 2005; Ballestar, 2010; Maiti et al., 2011). The same goes for other natural and artificial clones (Freund et al., 2013), as it is inevitably impossible to duplicate all the environmental interactions and internal bodily conditions involved in forming each individual phenotype (at the very least, the exact locations in space of each of a given pair of twins must permanently differ, causing slight differences in stimuli detection, and consequently also in the adaptive plasticity of the processes occurring in sensory areas). Such necessary individual differences, and the uniqueness of biological systems, enable us to account for things like intrinsic perspective, the first-personal point of view, and the self, by furnishing us with a natural source for these, and in fact this sort of naturalized conception of subjectivity does seem well justified empirically. Indeed, the major obstacle to creating artificial systems with subjective points of view may well be the combining of these complex biological factors responsible for influencing instances of such naturally formed perspectivalness. Does this necessarily mean that all artificial systems are immediately ruled out as potentially conscious (from-the-inside)? Is the ‘information’ utilized by artifacts just observer-relative information, as Searle (2013a,b) claims, being information just for us, whereas for artifacts like computers it is just one more piece of binary code being transformed into another? Such conclusions are certainly rather tempting, but our aim here is not to analyze such issues in any detail. We may confine ourselves to remarking that it does not seem fundamentally impossible for artificial systems to individuate information and form private informational states (O’Regan, 2012), though predicting whether it will ever be feasible to construct such systems (e.g., guided by similar biological principles of replication, adaptation, learning, etc.) seems like an issue more suited to futurology. At the same time, the ‘information’ processed by machines based on computational (von Neumann) architectures – even ones as complicated as modern supercomputers and expert systems – is rather more repeatable than individuated. The situation is somewhat different when artificial neural networks are taken into account (Ferber, 1999), though, as these to a certain extent mimic some biological mechanisms (e.g., parallel distributed processing, continuous learning, and adaptation mechanisms, etc.). Yet even such systems do not seem comfortably situated, or to want to fit in specific ecological niches, and neither have they adapted to some specific dynamically changing informational resources – they did not ever evolve so as to act in order to survive, reproduce, and so on, being designed instead just to simulate certain human or animal actions (e.g., face recognition, language production, decision making, etc.). If we still do not know all the necessary and sufficient ingredients of consciousness as a natural phenomenon, then certainly we know even less about the feasibility of any artificial or machine consciousness (Hollande, 2003; Torrance et al., 2007; Kurzweil, 2012; O’Regan, 2012).
Ultimately, however, it seems quite reasonable to restrict graded consciousness to those systems capable of individuating information and, on the basis of this, building subjective points of view. We shall therefore propose the following global limitation postulate: conscious experience can only be formed by unique information detectors, i.e., individual systems. It is likely that only dynamically changing biological or biologically based systems, evolved and situated somewhere on Earth (or some Earth-like ecosystem), meet these criteria. Earthly creatures (Millikan, 2004, p. 44) are able to form individuated informational states which may become conscious at a certain point in the processing that goes on. Our aim in the next section will therefore be to analyze this process more precisely, in the hope of establishing some relevant local constraints on graded consciousness.
Selection and Local Limitations
If, as was suggested earlier, not all information detected by a given organism and integrated within its specialized subsystems (e.g., primary sensory areas) is ultimately conscious, then there must be certain mechanisms according to which conscious states are selected. Many hypotheses and empirical findings concerning selection mechanisms at cognitive and neuronal levels have been put forward in science. One of the more promising correlations links consciousness with novel, non-standard or incongruent information (Dehaene and Naccache, 2001; Mudrik et al., 2014). In this context, conscious systems may be seen as specified difference detectors and conscious information may indeed be understood as ‘differences that make a difference’ – which is to concur with IIT. Roughly put, when the difference is ‘spotted,’ neuroplasticity switches on; new associations of neurons are formed and readied to discriminate similar states more efficiently in the future, increasing adaptation, survival rates, and so on. (Hence it is that learning and memory are often considered crucial functions for consciousness. See also Cleeremans (2011) and Frith (2011). Of course, not all discernible differences are equally likely to be detected: the system turns its attention to emotionally, contextually, or perceptually primed contents or stimuli (thus attention is also regarded as crucial – see Lamme, 2006). Yet according to Lamme (2006), the phenomenon of consciousness cannot be fully understood just in terms of one function or another: what we should rather be looking for is the specific neuronal mechanism responsible for generating consciousness.
Theoretical proposals and empirical models of neuronal selection most often explore the idea of multiple parallel processes distributed throughout the brain structures that selectively generate states of consciousness (in diverse configurations and different activity patterns). The proposals in question include the Global Workspace Theory (GWT; Baars, 1994; Baars et al., 2013), the Dynamic Core theory or theory of neuronal group selection (DC; Edelman and Tononi, 2000; Edelman, 2003; Edelman et al., 2011), the Recurrent Processing model (RP; Lamme, 2006), the Global Neuronal Workspace (GNW; Dehaene et al., 1998; Dehaene and Changeux, 2011), or Semantic Pointer Competition (SPC; Thagard and Steward, 2014). Borrowing some of their theoretical vocabulary, the process may roughly be described as follows (based on sensory perception): When the sensory information passes a certain detection threshold, action potentials in the relevant feedforward neuronal networks are fired, enabling one to discern basic features of the sensing scene (crucial for possible motor responses). Later on in the process, longer-lasting activations of higher and lower sensory areas result in the integration of particular aspects of the information into perceptual objects (i.e., localized re-entrant or recurrent processing). The neuronal activity may then propagate itself onto more distant corticothalamic regions, binding these together into wide-ranging firing clusters (semantic pointers, re-entrant loops) that may integrate sensory, motor, and executive fronto-parietal areas. The group that forms recurrent, synchronized activity inside such a DC ultimately wins the competition or ‘survives’ the process of neuronal selection, ‘governing’ the neuronal workspace for a given length of time (about 100–200 ms, in fact).
It is worth noting that the abovementioned models tend to imply a fairly late and sudden acquisition of consciousness occurring when some particular higher-order cortical activity occurs either more globally (GW, GNW, SPC) or less so (DC, RP), in the corticothalamic complex. However, they also reveal certain features that may incline one toward favoring the graded and extended view advocated here (ones also compatible, to a certain extent, with the ideas of IIT; see Tononi and Koch, 2014). For example, in the RP model proposed in Lamme (2006), recurrent activations that operate at a localized corticothalamic level may count as preconscious, or perhaps as conscious but not reportable. The primary consciousness conception introduced in Edelman (2003) and Edelman et al. (2011) goes even further in this direction, as it is supposed to be shared by most of the animal world and to precede all higher forms of consciousness (associated with re-entrant loops connecting certain memorized emotional values with current sensory inputs). Moreover, in what SPC proposes, consciousness is thought to be significantly extended globally, being associated there with species of gradually diminished complexity in respect of neuronal structure (Thagard and Steward, 2014, p. 79). However, the authors rather arbitrarily assume that fish are the least developed creature possessing consciousness). In addition, the biologically justified picture of consciousness described in recent versions of the GW model (Baars, 2005, 2012; Baars et al., 2013) is also significantly extended globally.
Are we likely to discover mechanisms responsible for selecting states of consciousness amongst some of the lower-level forms of neuronal activity, possibly in the primary sensory and motor areas? Is there any fundamental difference between widespread re-entrant connections and more localized recurrent, or even feedforward, firings – except, of course, for their varied spatiotemporal extent, directedness, frequencies, and/or underlying molecular mechanisms? The quantitative methods proposed by IIT may show which of the activity patterns differentiates and integrates more information (Tononi, 2008, 2010; Barrett and Seth, 2011), but since not all integrated information need be conscious (see above and Mudrik et al., 2014), another ingredient will be needed as a basis for the selection. At the same time, however, outside of the varied neuronal characteristics and functional results, we are rather unlikely to find that extra ingredient in the activation patterns themselves. Just putting the matter very roughly, it may be said that widespread re-entrant activations are most likely reportable (when connections reach prefrontal areas), and more localized recurrent processing will facilitate the formation of perceptual objects (percepts), whilst feedforward sweep ‘just’ enables the extraction of basic information necessary for fast motor responses (e.g., location, motion, shape, depth, and color). Nevertheless, the decision as to which of the mechanisms ultimately starts to ‘generate’ consciousness in its most rudimentary forms (as its true neural correlate or NCC; see Metzinger, 2000; Noë and Thompson, 2004; Lamme, 2006; Hohwy, 2009) seems always more or less arbitrary, being determined in due course by certain functional assumptions as to whether the subject is able to react appropriately, to differentiate, to learn, to choose, to decide, to recall, to report, etc.
Is it possible, we may ask, to point to any other, perhaps more general, non-neuronal process that might shape selection, avoiding at the same time any arbitrary conclusions about the exact cognitive function or neuronal mechanisms involved? In other words, what might the general difference be between informational and conscious states? The fundamental features of consciousness described earlier in the text, understood as abstract meta-characteristics (categories of more detailed features), are good candidates as starting points for this search. Conscious states, according to those four features, are dually accessible (subjectively and objectively), referential (about something), bodily determined (produced by the subject’s physical body interacting with the environment), and useful in action (pragmatically functional for a given subject). The question now is this: which of these features are shared by all informational states, and which constitute the differentia specifica of consciousness? According to the global limitation postulate, consciousness may be engendered only by systems able to individuate information. Therefore, all informational states relevant to consciousness must be individuated and subjective, though not all of them need ultimately be conscious. (This last point in turn raises the issue of unconscious subjectivity – a problem that calls for further analysis elsewhere; see Neisser, 2006). Conscious states, as selected-informational states, must share this constitutive ingredient; hence, all “consciousness is essentially subjective” (Searle, 1992, pp. 20–21; also Searle, 2000a, pp. 561–570). Both informational and conscious states are not just internally discernible for a given individual system, but also potentially visible to an external observer (e.g., as certain forms of neuronal activity or behavior): in this sense, both are dually (subjectively and objectively) accessible, even if not accessed at a given moment of time. Things are similar in the case of referentiality, since all the discernible differences that go to make up informational and conscious states must possess certain content, must be ‘about something’ if they are to be differentiated from one another. (Referential content may exhibit a sensory, emotional, perceptual, or any other sort of character – not necessarily one constitutive of any particular object). To the extent that all informational states are individuated in virtue of being detected by a specifically situated individual body engaged in unique environmental interactions, bodily determination should definitely also be considered a common feature. The final candidate for playing the role of a general differentiating element here would seem to be pragmatic functionality. The hypothesis that conscious states are selected from informational states by virtue of their usefulness in action certainly seems quite viable, so can pragmatic usability furnish us with a differentia specifica?
It is a trivial consequence of the mechanisms of evolution that a given creature will be better equipped to detect states if these resemble those that commonly proved useful for its ancestors. It is also evident from the perspective of developmental psychology and neuroscience that within biologically primed capacities for detecting this rather than that occurrence, once again, the most likely to be used are those states and action patterns resembling the ones that proved most useful during the relevant individual history. So we may assert that informational states are already pragmatically preselected by a given creature’s phylogenetic and ontogenetic history, and in that case it must also be rather trivially true to say that the ongoing selection process which ultimately leads to consciousness is pragmatically driven in general: i.e., that it is oriented toward selecting from the actions available to a given creature in response to moment-to-moment conditions. In this sense, the real, in vivo selecting of states to become gradually more conscious must be shaped by the actual situational context, emotional values, and internal/external conditions that a given system is in. We may add that the selection mechanisms indeed seem to operate as if their aim were to answer the question ‘What can I actually do, having this information?’ rather than the question ‘What do I know?’ (Here we might well wish to invoke such terms as affordances, contingencies, or expectations; see Gibson, 1977, 1979; O’Regan et al., 2006; Noë, 2006, 2010). Consequently, in this context, the assumption that the major role of consciousness is to govern or coordinate action at a given moment of time seems really quite viable, and in the light of this we can propose the following local limitation postulate: that conscious states are informational states selected to coordinate action at a given moment in time.
Consciousness, Information, and Action
Putting together the global and local limitation postulates just proposed, it seems that the aim of the present paper could after all be fulfilled, and an abstract concept of consciousness arrived at. This conception will concur with the general idea of IIT about the informational nature of conscious states and the graded nature of consciousness (Tononi and Koch, 2014). However, the extended range of graded forms of consciousness will be confined to individual systems (i.e., structurally unique, situated creatures, able to individuate information and adapt their actions to changing conditions), whereas differentiated informational states available within a given system are thought to undergo pragmatically driven, action-oriented, complex processes of selection and become, in due course, gradually more conscious. Consequently, consciousness may be characterized quite straightforwardly here: as a phenomenon in which a given individual system utilizes information in action. Put more succinctly, consciousness is individuated information utilized in action. Of course, such an abstract conception needs to be interpreted in the light of the distinctions and analyses developed here – ones that right from the very beginning have involved the concept of individuated information (characterized as evolutionarily embedded and/or socially altered and subjectively grounded) and a recognition of the fundamental features of consciousness (i.e., dual accessibility, referentiality, bodily determination, usefulness in action).
Translating the proposal into more practical terms, it may be said that conscious states consist of all possible sensory, perceptual, emotional, and other (kinds of) contents, selected from internally and/or externally induced informational states (differentiated and integrated in memory-based, sensory, or any other relevant subsystems of a given organism). The contents are selected to coordinate and adapt divergent actions to the specific conditions of the moment: they are utilized, for example, to grasp objects, to navigate motion, to see, to taste, to react appropriately, to decide, to reason, to think, to talk, to read, to answer questions, and so on. Conscious contents actually arise in the very acting process in which different components of actions (i.e., different interacting neuronal and bodily structures) show up as being, in a gradational sense, more or less conscious (leading in turn to a more global kind of task-related integration). Such interacting structures may refer, for example, to specific visual, auditory, tactile, olfactory, gustatory, or inner-bodily sensations, or to emotional feels, or to the location of certain objects in the surroundings, along with their usability, purpose or name, or to co-relations between things and other creatures, or similar past experiences, remembered events and situations, actual thoughts, ideas, concepts, words, phrases, meanings, beliefs, and so on. The degree to which the subject becomes aware of a given content (order of consciousness) is determined by a variety of heterogeneous factors, including general situational and emotional context, the applicability of a given action, the complexity and novelty of the informational states utilized in it, the spatiotemporal range of some given stimuli and the strength of the neuronal activations induced – to name just a few of the possibilities (Ramsøy and Overgaard, 2004; Overgaard et al., 2006, 2010; Mudrik et al., 2014).
Acting in a complex and dynamically changing environment, biological creatures are potentially able to differentiate between an infinite number of states; there can be no doubt that such overloaded informational conditions and such an ever-changing environment would have exerted evolutionary pressure, favoring more efficient selection mechanisms and greater plasticity in acting. The hypothesis that individuated information (with already preselected and subjective states) and consciousness (with action-specific state selection) are adaptive responses to those ‘needs’ seems plausible and consistent with the general scientific picture. The functional role of consciousness implied by this view may be described as that of providing online integration and refinement (adaptation) of available actions in accordance with the changing conditions of the moment. As far as we know, such pragmatically driven processes are most likely probabilistic and heuristic in nature (i.e., experience-based), and rather fallible or susceptible for error, yet nevertheless remain efficient enough to have persisted (Lotto and Purves, 2002; Rossano, 2003; Lotto, 2004). From this perspective, consciousness may be understood as yet another biological adaptation – a pragmatically functional process that enables the forming of unique patterns of action coordinated and refined on the basis of the private informational states available to a given agent.
Consequences and Conclusion
As was argued earlier, one of the characteristic features of conscious states is their referentiality (i.e., being about something). Some of the states, namely higher-order or metacognitive ones, refer to other, namely lower-order, states or contents. In most of the cases, however, a given subject is metacognitively unaware of the particular information utilized in action; economy and efficiency of acting dictate that most of the informational states will go metacognitively unnoticed, being impossible to report for the subject. For example, walking up a hill we are not aware of most of the complex equilibrioceptive information that our body uses in order to maintain balance (we just naturally incline into the slope). Perhaps the most controversial claim supported in this article is that even such forms of information, commonly regarded as genuinely unconscious, count as a rudimentary form of consciousness (namely sensorimotor consciousness) insofar as they are utilized for the purpose of acting. To put it simply, we are our bodies, so if our body ‘knows,’ then we know; in most cases it is just that we do not know that we know. One would probably argue that although such sensorimotor information is accessible to the subject, there is ‘nothing it is like to have it’ – no subjective experience in it, no qualia. There is insufficient space to analyze the complicated problem of qualia here (for that purpose, see Dennett, 1988; Crane, 2000; Bayne, 2009), but it is in fact subjectivity that is crucial in this context. The notion of individuated information developed in the present article thus furnishes at least a provisional answer to this objection. It proposes a naturalized conception of subjectivity itself – one based on the uniqueness of biological systems and their individual differences. Extending the analogy just deployed, it may thus be added that we are each of us not ‘any-body’: we are, rather, particular, individual bodies, so that no other body could ever ‘know’ anything in exactly the same way that our body does (in that all information accessible within the biological system is individuated). Ultimately, insofar as consciousness is ‘individuated information utilized in action’ by a given creature, what-it-is-like to have individuated information means what-it-is-like for a subject to act (where acting feels a certain way for the subject or is subjectively experienced as such).
In many philosophical theories (e.g., Rosenthal, 1986; Lycan, 1996; Gennaro, 2005) and dichotomous neurocognitive models (Dehaene et al., 2006; Dehaene and Changeux, 2011), consciousness is associated exclusively with certain higher-order metacognitive or reportable states. Except for certain problems and empirical shortcomings (mentioned earlier), such a position also brings the aim of the science of consciousness down to the level of a quest for a certain specific mechanism of generation or, at least, a contrastive threshold between higher-order conscious and lower-order unconscious states. The graded, action-oriented approach developed in this article offers a pragmatic explanation of why there may not be any contrastive difference between conscious and unconscious processes at the neuronal level (see Peremen and Lamy, 2014). The quest for a kind of ‘magic button’ that ‘switches the light on’ in some way or other may be altogether misleading, as the claim that consciousness does not arise suddenly is empirically well justified (as was shown above). According to the view supported here, consciousness shows up in the way in which living systems interact with the world using internal and external information (in this context, see also O’Regan and Noë, 2001; Noë and Thompson, 2004; Noë, 2006, 2010). It builds up gradually in the individual interacting system so that, at a certain point, the system may become able to report on the contents of its own states, yet this does not mean that it has just acquired consciousness while previously having been totally unconscious.
This gradational, action-oriented view also has certain methodological consequences. For example, it is definitely more difficult in experimental practice to find evidence for consciousness at lower levels of functioning – in this case, more sensitive measuring methods and tools are needed. If consciousness is indeed to be action-oriented, then perhaps certain quite new action-oriented measuring procedures will also need to be developed. At the same time, the gradational view does not call for a methodological revolution: for instance, contrastive analysis might also be useful here (when the contrast or threshold is properly defined and detectable). The change, rather, is above all meta-theoretical, because the thresholds or contrasts used in experiments are to be seen more as operational or practical than as definite. Other implications that could have a bearing on methodology might relate to the more holistic and externalistic approach needed when seeking to explain graded consciousness. A reasonable piece of advice in this respect would be to locate one’s methodology somewhere in between out-of-body experiences and locked-in syndrome (if, that is, we may be permitted to paraphrase the radically externalistic and internalistic methodologies by invoking these well-known states of impairment; see Jonkisz, 2012). It is probably enough just to go ‘out of our heads’ (Noë, 2010), toward the most important bodily determinants of consciousness, albeit without necessarily dismissing the fact of the ‘brain’s causal priority’ (Bickle, 2008, p. 272).
Certain consequences of the abstract concept of consciousness developed in this article might prove more significant viewed from a metaphysical perspective, but as detailed analysis is not feasible in this closing section, a few such implications will just be briefly indicated below. Consciousness, understood as ‘individuated information in action,’ is ubiquitous in the animal world: it is definitely not restricted to humans, or to mammals, as its lowest orders might be exhibited by virtually any living system able to make non-random use of environmental information in its actions. What is more, the phenomenon seems to be fundamentally divided epistemologically (because dually accessible), yet can still be, and hopefully is, ontologically unified. In the account developed here, consciousness may be seen as implied by or bound to the unique physical structure of any active system that individuates information.
Instead of offering further conclusions, let us end by listing a small number of questions and problems potentially raised by this conception. With such a broad notion of consciousness in play, does the traditional distinction between consciousness and unconsciousness still hold valid? Is it possible for a given creature to act without being conscious, or is acting always information-based, and hence to a certain degree conscious? Is it possible to invent certain criteria for ‘individuation’ in order, for example, to determine whether artificial systems could possibly ever individuate information? How can information be subjective (i.e., individuated) yet not conscious (i.e., not used in action)? What do we mean exactly by ‘action’? Questions such as these, prompted as they are by the conception of consciousness as ‘individuated information in action’ outlined here, certainly merit a discussion of their own.
Conflict of Interest Statement
The author declares that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Acknowledgments
This article was supported by an OPUS grant from the National Science Centre of Poland (UMO-2013/11/B/HS1/03968).
References
- Alvarez-Silva S., Alvarez-Silva I., Alvarez-Rodriguez J., Perez-Echeverria M. J., Campayo-Martinez A., Rodriguez-Fernandez F. L. (2006). Epileptic consciousness: concept and meaning of aura. Epilepsy Behav. 8 527–533. 10.1016/j.yebeh.2005.12.013 [DOI] [PubMed] [Google Scholar]
- Baars B. (1994). A thoroughly empirical approach to consciousness. Psyche 6 1–18. [Google Scholar]
- Baars B. (2005). Subjective experience is probably not limited to humans: the evidence from neurobiology and behavior. Conscious. Cogn. 14 7–21. 10.1016/j.concog.2004.11.002 [DOI] [PubMed] [Google Scholar]
- Baars B. (2012). The biological cost of consciousness. Nat. Proc. 1–16. 10.1038/npre.2012.6775.1 [DOI] [Google Scholar]
- Baars B., Franklin S., Ramsoy T. Z. (2013). Global workspace dynamics: cortical “binding and propagation” enables conscious contents. Front. Psychol. 4:200 10.3389/fpsyg.2013.00200 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Balconi M., Bortolotti A. (2013). Conscious and unconscious face recognition is improved by high-frequency rTMS on pre-motor cortex. Conscious. Cogn. 22 771–778. 10.1016/j.concog.2013.04.013 [DOI] [PubMed] [Google Scholar]
- Ballestar E. (2010). Epigenetics lessons from twins: prospects for autoimmune disease. Clin. Rev. Allergy. Immunol. 39 30–41. 10.1007/s12016-009-8168-4 [DOI] [PubMed] [Google Scholar]
- Barker R. G. (1968). Ecological Psychology: Concepts and Methods for Studying the Environment of Human Behavior. Stanford, CA: Stanford University Press. [Google Scholar]
- Barrett A. B., Seth A. K. (2011). Practical measures of integrated information for time-series data. PLoS Comput. Biol. 7:1 10.1371/journal.pcbi.1001052 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Bayne T. (2009). “Consciousness,” in The Routledge Companion to Philosophy of Psychology eds Symons J., Calvo P. (New York, NY: Routledge; ) 477–494. [Google Scholar]
- Bickle J. (2008). The molecules of social recognition memory: implications for social cognition, extended mind, and neuroethics. Conscious. Cogn. 2 468–474. 10.1016/j.concog.2008.03.015 [DOI] [PubMed] [Google Scholar]
- Block N. (1995). On confusion about a function of consciousness. Behav. Brain Sci. 2 227–287. 10.1017/S0140525X00038188 [DOI] [Google Scholar]
- Bogen J. E. (1995). On the neurophysiology of consciousness, an overview. Conscious. Cogn. 1 52–62. 10.1006/ccog.1995.1003 [DOI] [PubMed] [Google Scholar]
- Boly M., Seth A. K., Wilke M., Ingmundson P., Baars B., Laureys S., et al. (2013). Consciousness in humans and non-human animals: recent advances and future directions. Front. Psychol. 4:625 10.3389/fpsyg.2013.00625 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Brogaard B. (2011). Are there any unconscious perceptual processes? Conscious. Cogn. 20 449–463. 10.1016/j.concog.2010.10.002 [DOI] [PubMed] [Google Scholar]
- Brook A. (2008). Terminology in Consciousness Studies. Abstract Retrieved from ASSC Conference Abstracts Database. Ottawa, ON: University of Ottawa; Available at: http://www.ym.edu.tw/assc12/tutorials.html#02 (accessed December 12, 2014). [Google Scholar]
- Brook A., Raymont P. (2006). Preface to the Forthcoming Book: Unified Theory of Consciousness. Ottawa, ON: University of Ottawa; Available at: http://http-server.carleton.ca/~abrook/papers/2006-UnifiedConsc-Preface.pdf (accessed December 12, 2014). [Google Scholar]
- Carruthers P. (1998). Animal Subjectivity. Psyche 4:3 Available at: http://www.theassc.org/files/assc/2377.pdf (accessed December 12, 2014). [Google Scholar]
- Chalmers D. (1995). Facing up to the problem of consciousness. J. Conscious. Stud. 3 200–219. [Google Scholar]
- Chalmers D. (1996). The Conscious Mind: in Search of a Fundamental Theory. Oxford: Oxford University Press. [Google Scholar]
- Chalmers D. (2002). “Does conceivability entail possibility?,” in Conceivability and Possibility, eds Gendler T., Hawthorne J. (Oxford: Oxford University Press; ) 145–200. [Google Scholar]
- Chalmers D., Searle J. R. (1997). Consciousness and the Philosophers: An Exchange. New York: The New York Review of Books. Available at: http://www.nybooks.com/articles/archives/1997/may/15/consciousness-an-the-philosophers-an-exchange/ [Google Scholar]
- Clark A. (1997). Being There: Putting Brain, Body and World Together Again. Cambridge, MA: MIT Press. [Google Scholar]
- Clark A. (2008). Supersizing the Mind: Embodiment, Action, and Cognitive Extension. Oxford: Oxford University Press; 10.1093/acprof:oso/9780195333213.001.0001 [DOI] [Google Scholar]
- Cleeremans A. (1994). Awareness and abstraction are graded dimensions. Behav. Brain Sci. 17 402–403. 10.1017/S0140525X00035111 [DOI] [Google Scholar]
- Cleeremans A. (2011). The radical plasticity thesis: How the brain learns to be conscious. Front. Psychol. 2:86 10.3389/fpsyg.2011.00086 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Cleeremans A., Jiménez L. (2002). “Implicit learning and consciousness: a graded, dynamic perspective”, in Implicit Learning and Consciousness: An Empirical, Computational and Philosophical Consensus in the Making? eds French R. M., Cleeremans A. (Hove: Psychology Press; ) 1–40. [Google Scholar]
- Cohen M. A., Dennett D. C. (2011). Consciousness cannot be separated from function. Trends Cogn. Sci. 15 358–364. 10.1016/j.tics.2011.06.008 [DOI] [PubMed] [Google Scholar]
- Crane T. (2000) “The origins of qualia,” in The History of the Mind-Body Problem eds Crane T., Patterson S. (London: Routledge; ) 169–194. [Google Scholar]
- Crick F., Koch C. (2003). A framework for consciousness. Nat. Neurosci. 2 119–126. 10.1038/nn0203-119 [DOI] [PubMed] [Google Scholar]
- Damasio A. (1999). The Feeling of What Happens: Body, Emotion and the Making of Consciousness. London: Vintage. [Google Scholar]
- Dehaene S., Changeux J.-P. (2011). Experimental and theoretical approaches to conscious processing. Neuron 70 200–227. 10.1016/j.neuron.2011.03.018 [DOI] [PubMed] [Google Scholar]
- Dehaene S., Changeux J.-P., Naccache L., Sackur J., Sergent C. (2006). Conscious, preconscious, and subliminal processing: a testable taxonomy. Trends Cogn. Sci. 5 204–211. 10.1016/j.tics.2006.03.007 [DOI] [PubMed] [Google Scholar]
- Dehaene S., Kerszberg M., Changeux J. P. (1998). A neuronal model of a global workspace in effortful cognitive tasks. Proc. Natl. Acad. Sci. U.S.A. 95 14529–14534. 10.1073/pnas.95.24.14529 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Dehaene S., Naccache L. (2001). Towards a cognitive neuroscience of consciousness: basic evidence and a workspace framework. Cognition 79 1–37. 10.1016/S0010-0277(00)00123-2 [DOI] [PubMed] [Google Scholar]
- Dennett D. C. (1988). “Quining qualia,” in Consciousness in Modern Science eds Marcel A., Bisiach E. (Oxford: Oxford University Press; ) 121–158. [Google Scholar]
- Dennett D. C. (1991). Consciousness Explained. New York, NY: Little, Brown & Co. [Google Scholar]
- Dennett D. C. (1993). Review of searle, the rediscovery of the mind. J. Philos. 4 193–205. 10.2307/2940971 [DOI] [Google Scholar]
- Dennett D. C. (1995). The Mystery of Consciousness An Exchange. New York: New York Review of Books; Available at: http://www.nybooks.com/articles/archives/1995/dec/21/the-mystery-of-consciousness-an-exchange/ [Google Scholar]
- Dienes Z., Scott R. B. (2005). Measuring unconscious knowledge: distinguishing structural knowledge and judgment knowledge. Psychol. Res. 56 338–351. 10.1007/s00426-004-0208-3 [DOI] [PubMed] [Google Scholar]
- Edelman D., Seth A. (2009). Animal consciousness: a synthetic approach. Trends Neurosci. 9 476–484. 10.1016/j.tins.2009.05.008 [DOI] [PubMed] [Google Scholar]
- Edelman G. (2003). Naturalizing consciousness: a theoretical framework. Proc. Natl. Acad. Sci. U.S.A. 9 5520–5524. 10.1073/pnas.0931349100 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Edelman G., Gally J. A., Baars B. (2011). Biology of consciousness. Front. Psychol. 2:4 10.3389/fpsyg.2011.00004 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Edelman G., Tononi G. (2000). “Reentry and the dynamic core: neural correlates of conscious experience,” in Neural Correlates of Consciousness ed. Metzinger T. (Cambridge, MA: MIT Press; ) 139–151. [Google Scholar]
- Engel A. K., Maye A., Kurthen M., König P. (2013). Where’s the action? The pragmatic turn in cognitive science. Trends Cogn. Sci. 17 202–209. 10.1016/j.tics.2013.03.00 [DOI] [PubMed] [Google Scholar]
- Feinberg T. E., Mallatt J. (2013). The evolutionary and genetic origins of consciousness in the Cambrian Period over 500 million years ago. Front. Psychol. 4:667 10.3389/fpsyg.2013.00667 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Ferber J. (1999). Multi-Agent Systems. An Introduction to Distributed Artificial Intelligence. Reading, MA: Addison-Wesley. [Google Scholar]
- Fleming S. M., Weil R. S., Nagy Z., Dolan R. J., Rees G. (2010). Relating introspective accuracy to individual differences in brain structure. Science 5998 1541–1543. 10.1126/science.1191883 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Fraga M. (2005). From the cover: epigenetic differences arise during the lifetime of monozygotic twins. Proc. Natl. Acad. Sci. U.S.A. 102 10604–10609. 10.1073/pnas.0500398102 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Freund J., Brandmaier A. M., Lewejohann L., Kirste I., Kritzler M., Krüger A., et al. (2013). Emergence of individuality in genetically identical mice. Science 340 756–759. 10.1126/science.1235294 [DOI] [PubMed] [Google Scholar]
- Frith C. D. (2011). What brain plasticity reveals about the nature of consciousness: commentary. Front. Psychol. 2:87 10.3389/fpsyg.2011.00087 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Frith C. D., Perry R., Lumer E. (1999). The neural correlates of conscious experience: an experimental framework. Trends Cogn. Sci. 3 105–114. 10.1016/S1364-6613(99)01281-4 [DOI] [PubMed] [Google Scholar]
- Gennaro R. (2005). The HOT theory of consciousness: between a rock and a hard place. J. Conscious. Stud. 12 3–21. [Google Scholar]
- Giacino J. T. (2005). The minimally conscious state: defining the borders of consciousness. Prog. Brain Res. 150 381–95. 10.1016/S0079-6123(05)50027-X [DOI] [PubMed] [Google Scholar]
- Gibson J. (1977). “The theory of affordances,” in Perceiving, Acting, and Knowing eds Shaw R., Bransford J. (Hillsdale, NJ: Erlbaum; 19771) 67–82. [Google Scholar]
- Gibson J. (1979). The Ecological Approach to Visual Perception. Boston: Houghton Mifflin. [Google Scholar]
- Griffin D. R. (2001). Animal Minds: Beyond Cognition to Consciousness. Chicago, IL: University of Chicago Press. [Google Scholar]
- Griffin D. R., Speck G. B. (2004). New evidence of animal consciousness. Anim. Cogn. 7 5–18. 10.1007/s10071-003-0203-x [DOI] [PubMed] [Google Scholar]
- Hochstein S., Ahissar M. (2002). View from the top: hierarchies and reverse hierarchies in the visual system. Neuron 36 791–804. 10.1016/S0896-6273(02)01091-7 [DOI] [PubMed] [Google Scholar]
- Hohwy J. (2009). The neural correlates of consciousness. New experimental approaches needed? Conscious. Cogn. 18 428–438. 10.1016/j.concog.2009.02.006 [DOI] [PubMed] [Google Scholar]
- Holender D., Duscherer K. (2004). Unconscious perception: the need for a paradigm shift. Percept. Psychophys. 5 872–881. 10.3758/BF03194980 [DOI] [PubMed] [Google Scholar]
- Hollande O. (2003). Machine Consciousness. J. Conscious. Stud. 10 4–5. [Google Scholar]
- Hurley S. L. (1998). Consciousness in Action. Cambridge, MA: Harvard University Press. [Google Scholar]
- Hurley S. L. (2001). Perception and action: alternative views. Synthese 129 3–40. 10.1023/A:1012643006930 [DOI] [Google Scholar]
- Jackson F. (1982). Epiphenomenal qualia. Philos. Q. 32 127–136. 10.2307/2960077 [DOI] [Google Scholar]
- Jonkisz J. (2012). Consciousness: a four-fold taxonomy. J. Conscious. Stud. 19 55–82. [Google Scholar]
- Kihlstrom J. F. (1987). The cognitive unconscious. Science 237 1445–1452. 10.1126/science.3629249 [DOI] [PubMed] [Google Scholar]
- Kirk R. (1974). Zombies vs. Materialists. Proc. Aristotelian Soc. (Suppl. 48) 135–152. [Google Scholar]
- Kirk R. (2005). Zombies and Consciousness. Oxford: Clarendon Press; 10.1093/0199285489.001.0001 [DOI] [Google Scholar]
- Koch C. (2012). Consciousness – Confessions of a Romantic Reductionist. Cambridge, MA: MIT Press. [Google Scholar]
- Koch C., Tononi G. (2013). Reply to Searle in The New York Review of Books 3/2013 Can a Photodiode be Conscious? Available at: http://www.nybooks.com/articles/archives/2013/mar/07/can-photodiode-be-conscious/ [Google Scholar]
- Kokoszka A. (2007). States of Consciousness. Models for Psychology and Psychotherapy. New York: Springer. [Google Scholar]
- Kouider S., Gardelle V., Sackur J., Dupoux E. (2010). How rich is consciousness? Trends Cogn. Sci. 14 301–307. 10.1016/j.tics.2010.04.006 [DOI] [PubMed] [Google Scholar]
- Kouider S., Stahlhut C., Gelskov S. V., Barbosa L. S., Dutat M., deGardelle V., et al. (2013). A neural marker of perceptual consciousness in infants. Science 340 376–380. 10.1126/science.1232509 [DOI] [PubMed] [Google Scholar]
- Kriegel U. (2006). “Consciousness: phenomenal consciousness, access consciousness, and scientific practice,” in Handbook of Philosophy of Psychology and Cognitive Science ed. Thagard P. (Amsterdam: North-Holland; ) 195–217. [Google Scholar]
- Kriegel U. (2007). The same-order monitoring theory of consciousness. Synth. Philos. 2 361–384. [Google Scholar]
- Kurzweil R. (2012). How to Create a Mind: the Secret of Human Thought Revealed. New York: Viking. [Google Scholar]
- Lamme V. A. (2006). Towards a true neural stance on consciousness. Trends Cogn. Sci. 10 494–501. 10.1016/j.tics.2006.09.001 [DOI] [PubMed] [Google Scholar]
- Lamme V. A. (2010). How neuroscience will change our view on consciousness. Cogn. Neurosci. 3 204–220. 10.1080/17588921003731586 [DOI] [PubMed] [Google Scholar]
- Landry M., Appourchaux K., Raz A. (2014). Elucidating unconscious processing with instrumental hypnosis. Front. Psychol. 5:785 10.3389/fpsyg.2014.00785 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Långsjö J. W., Alkire M. T., Kaskinoro K., Hayama H., Maksimow A., Kaisti K. K., et al. (2012). Re-turning from oblivion: imaging the neural core of consciousness. J. Neurosci. 14 4935–4943. 10.1523/JNEUROSCI.4962-11.2012 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Legrain L., Cleermans A., Destrebecqz A. (2011). Distinguishing three levels in explicit self-awareness. Conscious. Cogn. 20 578–585. 10.1016/j.concog.2010.10.010 [DOI] [PubMed] [Google Scholar]
- Legrand D. (2007). Subjectivity and the body: introducing basic forms of self-consciousness. Conscious. Cogn. 16 577–582. 10.1016/j.concog.2007.06.011 [DOI] [PubMed] [Google Scholar]
- Levine J. (1983). Materialism and qualia: the explanatory gap. Pacific Philos. Q. 64 354–361. [Google Scholar]
- Levine J. (2001). Purple Haze: The Puzzle of Consciousness. Oxford: Oxford University Press; 10.1093/0195132351.001.0001 [DOI] [Google Scholar]
- Libet B. (1985). Unconscious cerebral initiative and the role of conscious will in voluntary action. Behav. Brain Sci. 8 529–566. 10.1017/S0140525X00044903 [DOI] [Google Scholar]
- Lotto B. (2004). Visual development: experience puts the colour in life. Curr. Biol. 14 619–621. 10.1016/j.cub.2004.07.045 [DOI] [PubMed] [Google Scholar]
- Lotto B., Purves D. (2002). The Empirical basis of color perception. Conscious. Cogn. 11 606–629. [DOI] [PubMed] [Google Scholar]
- Lycan W. G. (1996). Consciousness and Experience. Cambridge, MA: MIT Press. [Google Scholar]
- Lycan W. G. (2014). The intentionality of smell. Front. Psychol. 5:436 10.3389/fpsyg.2014.00436 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Lyvers M. (1999). Who has subjectivity? Psyche 5:31. [Google Scholar]
- Maiti S., Kumar K. H., Castellani C. A., O’Reilly R., Singh S. M. (2011). Ontogenetic de novo copy number variations (CNVs) as a source of genetic individuality: studies on two families with MZD twins for schizophrenia. PLoS ONE 6:e17125 10.1371/journal.pone.0017125 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Massoni S., Gajdos T., Vergnaud J.-C. (2014). Confidence measurement in the light of signal detection theory. Front. Psychol. 5:1455 10.3389/fpsyg.2014.01455 [DOI] [PMC free article] [PubMed] [Google Scholar]
- McGinn C. (1989). Can we solve the mind-body problem? Mind 98 349–366. 10.1093/mind/XCVIII.391.349 [DOI] [Google Scholar]
- Mealor A. D., Deines Z. (2013). The speed of metacognition: Taking time to get to know one’s structural knowledge. Conscious. Cogn. 22 123–136. 10.1016/j.concog.2012.11009 [DOI] [PubMed] [Google Scholar]
- Mele A. R. (2010). “Conscious deciding and the science of free will,” in Free Will and Consciousness, How Might they Work? eds Baumeister R. F., Mele A. R., Vohs K. D. (Oxford: Oxford University Press; ) 43–65. [Google Scholar]
- Merikle P. M., Reingold E. M. (1998). On demonstrating unconscious perception: comment on Draine and Greenwald. J. Exp. Psychol. Gen. 127 304–310. 10.1037/0096-3445.127.3.304 [DOI] [PubMed] [Google Scholar]
- Merikle P. M., Smilek D., Eastwood J. D. (2001). Perception without awareness: perspectives from cognitive psychology. Cognition 79 115–134. 10.1016/S0010-0277(00)00126-8 [DOI] [PubMed] [Google Scholar]
- Merker B. (2007). Consciousness without a cerebral cortex: a challenge for neuroscience and medicine. Behav. Brain Sci. 30 63–81. 10.1017/S0140525X07000891 [DOI] [PubMed] [Google Scholar]
- Metzinger T. (ed.) (2000). Neural Correlates of Consciousness: Empirical and Conceptual Questions (Cambridge, MA: The MIT Press; ). [Google Scholar]
- Migicovsky Z., Kovalchuk I. (2011). Epigenetic memory in mammals. Front. Gene. 2:28 10.3389/fgene.2011.00028 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Miller J., Schwarz W. (2014). Brain signals do not demonstrate unconscious decision making: an interpretation based on graded conscious awareness. Conscious. Cogn. 24 12–21. 10.1016/j.concog.2013.12.004 [DOI] [PubMed] [Google Scholar]
- Millikan R. (1989). Biosemantics. J. Philos. 86 281–297. 10.2307/2027123 [DOI] [Google Scholar]
- Millikan R. (2004). Varieties of Meaning. Cambridge, MA: MIT Press. [Google Scholar]
- Millikan R. (2006). “Useless content,” in Teleosemantics eds Macdonald G., Papineau D. (Oxford: Oxford University Press; ). [Google Scholar]
- Moller H. J., Devin G. M., Shen J., Shapiro C. M. (2006). Sleepiness is not the inverse of alertness: evidence from four sleep disorder patient groups. Exp. Brain Res. 173 258–266. 10.1007/s00221-006-0436-4 [DOI] [PubMed] [Google Scholar]
- Morin A. (2006). Levels of consciousness and self-awareness. Conscious. Cogn. 15 358–371. 10.1016/j.concog.2005.09.006 [DOI] [PubMed] [Google Scholar]
- Mudrik L., Faivre N., Koch C. (2014). Information integration without awareness. Trends Cogn. Sci. 18 488–496. 10.1016/j.tics.2014.04.009 [DOI] [PubMed] [Google Scholar]
- Nagel T. (1974). What is it like to be a bat? Philos. Rev. 83 435–451. 10.2307/2183914 [DOI] [Google Scholar]
- Natsoulas T. (1983). Concepts of consciousness. J. Mind Behav. 1 13–59. [Google Scholar]
- Natsoulas T. (1997a). Consciousness and self-awareness: Part I. Consciousness1, consciousness2, and consciousness3. J. Mind Behav. 18 53–74. [Google Scholar]
- Natsoulas T. (1997b). Consciousness and self-awareness: Part II. Consciousness4, consciousness5, and consciousness6. J. Mind Behav. 18 53–74. [Google Scholar]
- Neisser U. (1988). Five kinds of self-knowledge. Philos. Psychol. 1 35–59. 10.1080/09515088808572924 [DOI] [Google Scholar]
- Neisser U. (2006). Unconscious subjectivity. Psyche 12:3 Available at: http://www.theassc.org/files/assc/2642.pdf [accessed 20 January, 2015]. [Google Scholar]
- Nieuwenhuis S., de Kleijn R. (2011). Consciousness of targets during the attentional blink: a gradual or all-or-none dimension? Attent. Percept. Psychophys. 73:2 10.3758/s13414-010-0026-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Noë A. (2006). Precis of action in perception. Psyche. Available at: http://www.theassc.org/files/assc/2624.pdf (accessed December 12, 2014). [Google Scholar]
- Noë A. (2010). Out of Our Heads, Why You Are Not Your Brain, and Other Lessons from the Biology of Consciousness. New York, NY: Hill and Wang. [Google Scholar]
- Noë A., Thompson E. (2004). Are there neural correlates of consciousness? J. Conscious. Stud. 1 3–28. [Google Scholar]
- Northoff G., Musholt K. (2006). How can Searle avoid property dualism? Epistemic-ontological inference and autoepistemic limitation. Philos. Psychol. 5 1–17. 10.1080/09515080600901889 [DOI] [Google Scholar]
- O’Regan J. K. (2012). How to build a robot that is conscious and feels. Minds Mach. 22 117–136. 10.1007/s11023-012-9279-x [DOI] [Google Scholar]
- O’Regan J. K., Myin E., Noë A. (2006). Skill, corporality and alerting capacity in an account of sensory consciousness. Prog. Brain Res. 150 55–68. 10.1016/S0079-6123(05)50005-0 [DOI] [PubMed] [Google Scholar]
- O’Regan J. K., Noë A. (2001). A sensorimotor account of vision and visual consciousness. Behav. Brain Sci. 5 883–917. 10.1017/s0140525x01000115 [DOI] [PubMed] [Google Scholar]
- Overgaard M. (2004). Confounding factors in contrastive analysis. Synthese 141 217–231. 10.1023/B:SYNT.0000043019.64052.e0 [DOI] [Google Scholar]
- Overgaard M., Fehl K., Mouridsen K., Bergholt B., Cleeremans A. (2008). Seeing without seeing? Degraded conscious vision in a blindsight patient. PLoS ONE 3:3028 10.1371/journal.pone.0003028 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Overgaard M., Rote J., Mouridsen K., Ramsoy T. Z. (2006). Is conscious perception gradual or dichotomous? A comparison of report methodologies during a visual task. Conscious. Cogn. 15 700–708. 10.1016/j.concog.2006.04.002 [DOI] [PubMed] [Google Scholar]
- Overgaard M., Sandberg K. (2012). Kinds of access: different methods for report reveal different kinds of metacognitive access. Philos. Trans. R. Soc. B 367 1287–1296. 10.1098/rstb.2011.0425 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Overgaard M., Timmermans B., Sandberg K., Cleeremans A. (2010). Optimizing subjective measures of consciousness. Conscious. Cogn. 19 682–684. 10.1016/j.concog.2009.12.018 [DOI] [PubMed] [Google Scholar]
- Palmer T. D., Ramsey A. K. (2012). The function of consciousness in multisensory integration, Cognition 125 353–364. 10.1016/j.cognition.2012.08.003 [DOI] [PubMed] [Google Scholar]
- Panagiotaropoulos T. I., Deco G., Kapoor V., Logothetis L. K. (2012). Neuronal discharges and gamma oscillations explicitly reflect visual consciousness in the lateral prefrontal cortex. Neuron 5 924–935. 10.1016/j.neuron.2012.04.013 [DOI] [PubMed] [Google Scholar]
- Peremen Z., Lamy D. (2014). Do conscious perception and unconscious processing rely on independent mechanisms? A meta-contrast study. Conscious. Cogn. 24 22–32. 10.1016/j.concog.2013.12.006 [DOI] [PubMed] [Google Scholar]
- Pfefferbaum A., Sullivan E. V., Carmelli D. (2004). Morphological changes in aging brain structures are differentially affected by time-linked environmental influences despite strong genetic stability. Neurobiol. Aging 25 175–183. 10.1016/S0197-4580(03)00045-9 [DOI] [PubMed] [Google Scholar]
- Philippi C., Feinstein J. S., Khalsa S. S., Damasio A., Travel D., Landing G., et al. (2012). Preserved self-awareness following extensive bilateral brain damage to the insula, anterior cingulate, and medial prefrontal cornices. PLoS ONE 8:e38413 10.1371/journal.pone.0038413 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Pierre J. (2003). “Intentionality,” in Stanford Encyclopedia of Philosophy ed. Zalta E. N. (Stanford: Stanford University; ). Available at: http://plato.stanford.edu/entries/intentionality/#9 [Google Scholar]
- Ramsøy T. Z., Overgaard M. (2004). Introspection and subliminal perception. Phenomenol. Cogn. Sci. 3 1–23. 10.1023/B:PHEN.0000041900.30172.e8 [DOI] [Google Scholar]
- Reiss D., Marino L. (2001). Self-recognition in the bottlenose dolphin: a case of cognitive convergence. Proc. Natl. Acad. Sci. U.S.A. 10 5937–5942. 10.1073/pnas.101086398 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Ro T., Shelton D., Lee O. L., Chang E. (2004). Extrageniculate mediation of unconscious vision in transcranial magnetic stimulation-induced blindsight. Proc. Natl. Acad. Sci. U.S.A. 101 9933–9935. 10.1073/pnas.0403061101 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Rosenthal D. (1986). Two concepts of consciousness. Philos. Stud. 49 329–359. 10.1007/BF00355521 [DOI] [Google Scholar]
- Rossano M. J. (2003). Expertise and the evolution of consciousness. Cognition 89 207–36. 10.1016/S0010-0277(03)00120-3 [DOI] [PubMed] [Google Scholar]
- Saalmann Y. B. (2014). Intralaminar and medial thalamic influence on cortical synchrony, information transmission and cognition. Front. Syst. Neurosci. 8 83 10.3389/fnsys.2014.00083 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Sandberg K., Bobby B. M., Timmermans B., Cleeremans A., Overgaard M. (2011). Measuring consciousness: task accuracy and awareness as sigmoid functions of stimulus duration. Conscious. Cogn. 20 1659–1675. 10.1016/j.concog.2011.09.002 [DOI] [PubMed] [Google Scholar]
- Sandberg K., Timmermans B., Overgaard M., Cleeremans A. (2010). Measuring consciousness: is one measure better than the other? Conscious. Cogn. 4 1069–1078. 10.1016/j.concog.2009.12.013 [DOI] [PubMed] [Google Scholar]
- Schnakers C. (2008). A French validation study of the Coma Recovery Scale-Revised (CRS-R). Brain Injury 22 786–792. 10.1080/02699050802403557 [DOI] [PubMed] [Google Scholar]
- Scott R. B., Dienes Z. (2008). The conscious, the unconscious, and familiarity. J. Exp. Psychol. Learn. Mem. Cogn. 34 1264–1288. 10.1037/a0012943 [DOI] [PubMed] [Google Scholar]
- Scott R. B., Dienes Z. (2010). Knowledge applied to new domains: the unconscious succeeds where the conscious fails. Conscious. Cogn. 19 391–398. 10.1016/j.concog.2009.11.009 [DOI] [PubMed] [Google Scholar]
- Searle J. (1992). The Rediscovery of the Mind. Cambridge, MA: MIT Press. [Google Scholar]
- Searle J. (1995). The Mystery of Consciousness. An Exchange. New York: The New York Review of Books; Available at: http://www.nybooks.com/articles/archives/1995/dec/21/the-mystery-of-consciousness-an-exchange/ (accessed December 12, 2014). [Google Scholar]
- Searle J. (1997). Consciousness and the Philosophers. New York: The New York Review of Books; Available at: http://www.nybooks.com/articles/archives/1997/mar/06/consciousness-the-philosophers/ [Google Scholar]
- Searle J. (2000a). Consciousness. Annu. Rev. Neurosci. 23 557–578. 10.1146/annurev.neuro.23.1.557 [DOI] [PubMed] [Google Scholar]
- Searle J. (2000b). Consciousness, free action and the brain. J. Conscious. Stud. 7 3–22. [Google Scholar]
- Searle J. (2013a). Can Information Theory Explain Consciousness. New York: The New York Review of Books; Available at: http://www.nybooks.com/articles/archives/2013/jan/10/can-information-theory-explain-consciousness/ [Google Scholar]
- Searle J. (2013b). Can a Photodiode be Conscious. New York: The New York Review of Books. Available at: http://www.nybooks.com/articles/archives/2013/mar/07/can-photodiode-be-conscious/ [Google Scholar]
- Sergent C., Dehaene S. (2004). Is consciousness a gradual phenomenon? Evidence for an all-or none bifurcation during the attentional blink. Psychol. Sci. 11 720–729. 10.1111/j.0956-7976.2004.00748.x [DOI] [PubMed] [Google Scholar]
- Seth A. K., Baars B., Edelman D. (2005). Criteria for consciousness in humans and other mammals. Conscious. Cogn. 1 119–139. 10.1016/j.concog.2004.08.006 [DOI] [PubMed] [Google Scholar]
- Seth A. K., Dienes Z., Cleeremans A., Overgaard M., Pessoa L. (2008). Measuring consciousness: relating behavioral and neurophysiological approaches. Trends Cogn. Sci. 12 314–321. 10.1016/j.tics.2008.04.008 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Shea N., Boldt A., Bang D., Yeung N., Heyes C., Frith C. D. (2014). Supra-personal cognitive control and metacognition. Trends Cogn. Sci. 18:1 10.1016/j.tics.2014.01006 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Siewert C. (2006). “Consciousness and intentionality,” in Stanford Encyclopedia of Philosophy ed. Zalta E. N. (Stanford: Stanford University; ). Available at: http://plato.stanford.edu/entries/consciousness-intentionality/ [Google Scholar]
- Smith J. D. (2009). The study of animal metacognition. Trends Cogn. Sci. 9 389–396. 10.1016/j.tics.2009.06.009 [DOI] [PubMed] [Google Scholar]
- Soon C. S., Brass M., Heinze H. J., Haynes J. D. (2008). Unconscious determinants of free decisions in the human brain. Nat. Neurosci. 11 543–545. 10.1038/nn.2112 [DOI] [PubMed] [Google Scholar]
- Stoljar D. (2007). Two conceivability arguments compared. Proc. Aristotelian Soc. 108 27–44. 10.1111/j.1467-9264.2007.00210.x [DOI] [Google Scholar]
- Stuss D. T., Anderson V. (2004). The frontal lobes and theory of mind: developmental concepts from adult focal lesion research. Brain Cogn. 55 69–83. 10.1016/S0278-2626(03)00271-9 [DOI] [PubMed] [Google Scholar]
- Swaddle J. P., Cathey M. G., Cornell M., Hopkinton B. P. (2005). Socially transmitted mate preferences in a monogamous bird: a non-genetic mechanism of sexual selection. Proc. Biol. Sci. R. Soc. 272:1567 10.1098/rspb.2005.3054 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Szczepanowski R., Traczyk J., Wierzchoñ M., Cleeremans A. (2013). The perception of visual emotion: comparing different measures of awareness. Conscious Cogn. 22 212–220. 10.1016/j.concog.2012.12.003 [DOI] [PubMed] [Google Scholar]
- Tart C. (1969). Altered States of Consciousness: a Book of Readings. New York: Wiley [Google Scholar]
- Teasdale G., Jennett B. (1974). Assessment of coma and impaired consciousness. A practical scale. Lancet II 81–86. 10.1016/S0140-6736(74)91639-0 [DOI] [PubMed] [Google Scholar]
- Thagard P., Steward T. C. (2014). Two theories of consciousness: semantic pointer competition vs. Inform. Itegr. Conscious. Cogn. 30 73–90. 10.1016/j.concog.2014.07.001 [DOI] [PubMed] [Google Scholar]
- Thompson E., Varela F. (2001). Radical embodiment: neural dynamics and consciousness. Trends Cogn. Sci. 5 418–425. 10.1016/S1364-6613(00)01750-2 [DOI] [PubMed] [Google Scholar]
- Tononi G. (2004). An information integration theory of consciousness. BMC Neurosci. 5:42 10.1186/1471-2202-5-42 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Tononi G. (2008). Consciousness as integrated information: a provisional manifesto. Biol. Bull. 215 216–242. 10.2307/25470707 [DOI] [PubMed] [Google Scholar]
- Tononi G. (2010). Information integration: its relevance to brain function and consciousness. Arch. Ital. Biol. 148 299–322. [PubMed] [Google Scholar]
- Tononi G., Koch C. (2014). Consciousness: here, there but not everywhere? Philos. Trans. R. Soc. Lond. B Biol. Sci. 370 pii: 20140167 10.1098/rstb.2014.0167 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Torrance S., Clowes R., Chrisley R. (eds) (2007). Machine consciousness embodiment and imagination. J. Conscious. Stud. 14:7. [Google Scholar]
- van Gaal S., Lamme V. A. F. (2012). Unconscious high-level information processing: implication for neurobiological theories of consciousness. Neuroscientist 18 287–301. 10.1177/1073858411404079 [DOI] [PubMed] [Google Scholar]
- van Gaal S., Lamme V. A., Fahrenfort J. J., Ridderinkhof K. R. (2011). Dissociable brain mechanisms underlying the conscious and unconscious control of behaviour. J. Cogn. Neurosci. 23 91–105. 10.1162/jocn.2010.21431 [DOI] [PubMed] [Google Scholar]
- Varela F. J., Thompson E., Rosch E. (1991). The Embodied Mind. Cambridge, MA: MIT Press. [Google Scholar]
- Wierzchoñ M., Paulewicz B., Asanowicz D., Timmerman B., Cleeremans A. (2014). Different subjective awareness measures demonstrate the influence of visual identification on perceptual awareness ratings. Conscious. Cogn. 27 109–120. 10.1016/j.concog.2014.04.009 [DOI] [PubMed] [Google Scholar]
- Windey B., Gevers W., Cleeremans A. (2013). Subjective visibility depends on level of processing, Cognition 2 404–409. 10.1016/j.cognition.2013.07.012 [DOI] [PubMed] [Google Scholar]
- Windey B., Vermeiren A., Atas A., Cleeremans A. (2014). The graded and dichotomous nature of visual awareness. Philos. Trans. R. Soc. Lond. Ser. B Biol. Sci. 369:1641 10.1098/rstb.2013.0282 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Wittgenstein L. (2009). Philosophical Investigations 4th Edn transl. eds Anscombe G. E. M., Hacker P. M. S., Schulte J. (Oxford: Blackwell; ). [Google Scholar]
- Zelazo P. D. (2004). The development of conscious control in childhood. Trends Cogn. Sci. 8 12–17. 10.1016/j.tics.2003.11.001 [DOI] [PubMed] [Google Scholar]
