My aim in this article is to examine the field of consciousness research with an analysis of 78 of the most important papers in the field. In doing so, I compare the empirical foundations these disciplines rely upon and detail how each approaches the measurement of conscious experience.

There is no consensus on how to define consciousness, and thus no single discipline dedicated to its study. Instead, thinkers across all fields have developed separate theoretical frameworks and thought experiments, which often remain isolated. To better understand consciousness research and find clusters of insights, I built an instrument to employ large language models—o1, Sonnet, and Gemini—to categorize any paper selected, and extract 20 key facts from it. These facts are supported by evidence and context from the paper using a recursive prompting technique designed to generate accurate and structured output. The instrument then automatically creates a vector representation of the facts using OpenAI embeddings, which mathematically identify both interdisciplinary and intra-disciplinary links between the facts. This allows me to retrieve connections across disciplines, with embedding search algorithms, semantic similarity analysis, and an interactive UMAP dimensionality reduction based visualization that turns the 3072 dimension vectors into a 3D map.

 

A centrally positioned cluster is that of the “Hard Problem of Consciousness”, which was introduced by David Chalmers In 1996. He posed a fundamental question: “How could a physical system such as a brain also be an experiencer?”. Through my examination of the primary disciplines involved in consciousness research – Neuroscience, Philosophy of Mind, Quantum Theories of Consciousness, Artificial Intelligence and Computational models – I found that the “Hard Problem” was one of the few ideas explored across each of them having been discussed by prominent researchers such as DeepMind Senior Engineer Murray Shanahan, Nobel Prize winner Francis Crick, and MIT physics professor Max Tegmark.

Red: Neuroscience

Blue: AI

Green: Quantum

Purple: Philosophy of Mind

The one idea connecting all of consciousness research is that consciousness research is hard.

This connection point is also a point of contention, with some entertaining it as a thought experiment, most notably in AI and Philosophy of Mind, and some responding to it with criticism. A notable retort came from Salk Institute Neurophilosopher Patricia Churchland, who urged further exploration of the fundamental functions of the nervous system, comparing the “hard problem of consciousness” to any other neurobiological puzzle.

Patricia Churchland’s approach exemplifies the empirical methodology of neuroscience, focusing on measuring consciousness rather than exploring it through thought experiments. Churchland’s Salk Institute colleague and winner of the 1962 Nobel Prize in Medicine, Francis Crick, spent much of his career attempting to solve the neurobiological problem of consciousness. As an undergrad research intern at the Salk Institute, I was introduced to Crick’s concept of the Neurobiological Correlates of Consciousness (NCC), the brain activity patterns directly associated with consciousness.

This study of the NCC has pushed our understanding of consciousness past armchair speculation, enabling empirical investigation through techniques such as electroencephalography (EEG). This approach allows us to explore the role of synchronized neural oscillations in the beta and gamma frequency bands where studies suggest that synchrony of large neural oscillations (brainwaves) may be critical for the processing of diverse perceptual features such as color, shape and motion into a unified conscious percept. They propose that consciousness arises when spatially distributed neurons fire in synchrony.

 

Red: neurobiology

It can be observed that synchronized or orderly action is a core idea in the biological studies of consciousness, bridging a number of the major theories. The observations that synchronised firing of neurons are associated with consciousness lend support to the Entropic Brain Hypothesis (EBH), which suggests that waking consciousness promotes orderly states within the brain, reducing entropy. Although formally introduced by psychopharmacologist Robin Carhart-Harris in 2014 based on his studies on psychedelics, EBH may trace its roots to Erwin Schrödinger’s 1944 contention that reducing entropy was the object of life. Within the cluster of facts related to synchrony and order, the experimental evidence includes TMS and EEG research demonstrating how synchronized neuronal firing can reflect lower-entropy states, studies indicating that conscious and unconscious brain states can be distinguished and investigations into psychedelic compounds that indicate an increase brain entropy. This is shown by a rise of signal variance in neuroimaging studies, leading to perceived alterations in ego and conscious experience.

 

Adding to neural synchrony and negative entropy are studies that examine feedback mechanisms in conscious perception. Researchers have demonstrated that an initial feedforward sweep of sensory information can register details without conscious awareness; however, the presence of feedback from higher-order areas to sensory regions of the brain leads to conscious experiences. Consistent with this view, Global Workspace Theory (GWT) highlights how specific brain networks become active when a stimulus achieves conscious access. The activation of these brain regions effectively broadcasts information across the brain, enabling the cognitive processes associated with consciousness in such as introspection, planning, and verbal report.

These theories demonstrate an empirical approach to the study of consciousness. They operate on a reasonable assumption that humans are conscious, and approach the study from a human biological perspective. They, however, need to be combined with the other disciplines to discover the true nature of consciousness. To measure the exchange of ideas between the fields, I computed the mean vector of the embeddings in each field and found their midpoint. I then measured the similarity scores between each field:

The similarity between AI and Cognitive and Biological Sciences is 0.8 is equal to that between AI and philosophy of mind. This provides an interesting insight into the approach AI consciousness researchers have taken to the discipline. The question of whether AI can be conscious has become a matter of debate. Patrick Butlin and Robert Long’s seminal paper, Consciousness in Artificial Intelligence: Insights from the Science of Consciousness, provides a comprehensive survey of consciousness research from numerous theoretical perspectives and concludes that current AI technologies are unlikely to be conscious, but note it is not impossible, in principle, for an artificial system to achieve consciousness.

This conclusion has ushered in some empirical research but the field of AI consciousness remains largely theoretical, characterized by thought experiments and conceptual frameworks rather than concrete investigations into the Artificial Correlates of Consciousness (ACC). The study of information processing inside neural networks is itself an emerging field, making the investigation of AI consciousness even more novel. Nevertheless, these early theoretical foundations and research directions will likely shape the field's future development in significant ways.

A December 2024 publication in Nature provides an empirical framework and approach to testing whether GPT-3 might exhibit consciousness by administering adapted cognitive and emotional intelligence tests. This paper, instead of discussing the opposing positions of Searle and Wittgenstein on language, operated under the assumption that the meaning of the word is an activity, rejected the Chinese Room (Computer can write Chinese but doesn't understand) argument and tested whether GPT-3 might exhibit consciousness by administering adapted cognitive and emotional intelligence tests.

This method highlights a shift in approach, moving towards a more human view of AI systems and their study. Murray Shanahan, in his paper Simulacra as Conscious Exotica critiqued an anthropomorphised view of AI due to concerns of misrepresentation of their capabilities and understanding yet I find the distinction to be dramatized. An example being my professor Terrence Sejnowski’s 2024 paper Transformers and cortical waves: encoders for pulling in context across time which describes a transformer-like architecture in the human brain demonstrating an attention-like mechanism.

Exploring the similarities and differences between humans and LLMs can illuminate unexplored areas of consciousness or lead to new capabilities in AI systems. Drawing on neurobiology’s fundamental assumption that other humans are conscious, I speculate that these investigative strategies could be applied to the study of consciousness in artificial systems. Low similarity(0.7) between AI and the physical theories of consciousness also represent a potential area for further research potentially building on findings that THz radiation and ultrasound can alter membrane permeability and affect consciousness in humans as well as semiconductor logic calculations. Despite being a young field, AI consciousness research could help connect the main disciplines in the space, offering a computational approach to consciousness research.

"Computational models are essential for unraveling the complexities of the brain, allowing us to simulate and understand neural mechanisms in ways that purely experimental approaches cannot."

Terrence Sejnowski

5

0
0

Reactions

0
0

More posts like this

Comments1
Sorted by Click to highlight new comments since:

Executive summary: This post examines consciousness research across disciplines, emphasizing the lack of consensus on defining consciousness, the empirical advancements in neuroscience, and the theoretical exploration in AI, while highlighting the interconnected nature of these studies and the potential for AI consciousness research to bridge these fields. Key points:

  1. There is no consensus on a definition of consciousness, and its study spans multiple disciplines—neuroscience, philosophy of mind, quantum theories, and AI—with limited integration between them.
  2. The "Hard Problem of Consciousness," introduced by David Chalmers, is a central and contentious topic discussed across all disciplines, contrasting empirical neuroscience with more theoretical approaches in AI and philosophy.
  3. Neuroscience focuses on empirical methods, such as studying the Neurobiological Correlates of Consciousness (NCC) and neural synchrony, which reveal connections between brain activity and conscious experiences.
  4. AI consciousness research remains largely theoretical, with foundational studies exploring the potential for artificial systems to achieve consciousness, though current AI technologies are considered unlikely to be conscious.
  5. Interdisciplinary studies reveal significant overlap between AI and cognitive sciences (similarity score: 0.8), suggesting potential for AI research to inform consciousness studies, particularly through computational models.
  6. Emerging approaches, such as adapted cognitive and emotional intelligence tests for AI systems like GPT-3, signal a shift toward human-centric studies of AI consciousness, with potential implications for broader understanding of consciousness.

 

 

This comment was auto-generated by the EA Forum Team. Feel free to point out issues with this summary by replying to the comment, and contact us if you have feedback.

Curated and popular this week
Relevant opportunities