GNSS & Machine Learning Engineer

Category: Philosophy

GPT-4 in the top 1% of human thinkers in creativity test

In a recent study by the University of Montana, GPT-4 demonstrated remarkable performance in the Torrance Tests of Creative Thinking (TTCT, a standard test for measuring creativity), matching the top 1% of human thinkers. The model excelled in fluency and originality. These findings imply that the creative abilities of GPT-4 could potentially surpass those of humans.

For a recent benchmark on advanced reasoning capabilities of large language models take a look at the ARB (Advanced Reasoning Benchmark).

Comments on Common AI Questions

The field of artificial intelligence raises numerous questions, frequently discussed but often left without a clear consensus. We’ve chosen to contribute our unique insights on some of these recurring topics, aiming to shed light on them from our perspective. In formulating this text, we’ve utilized GPT-4 to assist with language generation, but the insights and conclusions drawn are entirely our own.

The questions we address are:

  • Can Machines Develop Consciousness?
  • Should Humans Verify Statements from Large Language Models (LLMs)?
  • Can Large Language Models (LLMs) Generate New Knowledge?

From the philosophical implications to practical applications, these topics encompass the broad scope of AI’s capabilities and potential.

Can Machines Develop Consciousness? A Subjective Approach

The question of whether machines can develop consciousness has sparked much debate and speculation. A fruitful approach might be to focus not solely on the machines themselves, but also on our subjective interpretations and their influence on our understanding of consciousness.

Consciousness might not be directly definable, but its implications could be essential for our predictive abilities. When we assign consciousness to an entity – including ourselves – we could potentially enhance our ability to anticipate and understand its behavior.

Associated attributes often assigned to consciousness include self-reflection, self-perception, emotional experience, and notably, the capacity to experience pain, as highlighted by historian Yuval Noah Harari. However, recognizing these attributes in an object is a subjective process. It is not inherently possessed by the object but is a projection from us, the observers, based on our interpretation of the object’s behaviors and characteristics.

This suggests that a machine could be considered “conscious” if assigning such traits improves our understanding and prediction of its behavior. Interestingly, this notion of consciousness assignment aligns with a utilitarian perspective, prioritizing practicality and usefulness over abstract definitions.

Reflecting on consciousness might not always be a conscious and rationalized process. Often, our feelings and intuition guide us in understanding and interpreting the behaviors of others, including machines. Therefore, our subconscious might play a crucial role in determining whether we assign consciousness to machines. In this light, it might make sense to take a democratic approach in which individuals report their feelings or intuitions about a machine, collectively contributing to the decision of whether to assign it consciousness.

Furthermore, reflexivity, commonly associated with consciousness, could potentially be replicated in machines through a form of “metacognitive” program. This program would analyze and interpret the output of a machine learning model, mirroring aspects of self-reflection (as in SelFee). Yet, whether we choose to perceive this program as part of the same entity as the model or as a separate entity may again depend on our subjective judgment.

In conclusion, the concept of consciousness emerges more from our personal perspectives and interpretations than from any inherent qualities of the machines themselves. Therefore, determining whether a machine is ‘conscious’ or not may be best decided by this proposed democratic process. The crucial consideration, which underscores the utilitarian nature of this discussion, is that attributing consciousness to machines could increase our own predictive abilities, making our interactions with them more intuitive and efficient. Thus, the original question ‘Can machines develop consciousness?’ could be more usefully reframed as ‘Does it enhance our predictability, or feel intuitively right, to assign consciousness to machines?’ This shift in questioning underscores the fundamentally subjective and pragmatic nature of this discussion, engaging both our cognitive processes and emotional intuition.

Should Humans Verify Statements from Large Language Models (LLMs)? A Case for Autonomous Verification

In the realm of artificial intelligence (AI), there is ongoing discourse regarding the necessity for human verification of outputs generated by large language models (LLMs) due to their occasional “hallucinations”, or generation of factually incorrect statements. However, this conversation may need to pivot, taking into account advanced and automated verification mechanisms.

LLMs operate by predicting the most probable text completion based on the provided context. While they’re often accurate, there are specific circumstances where they generate “hallucinations”. This typically occurs when the LLM is dealing with a context where learned facts are absent or irrelevant, leaving the model to generate a text completion that appears factually correct (given its formal structure), but is indeed a fabricated statement. This divergence from factuality suggests a need for verification, but it doesn’t inherently demand human intervention.

Rather than leaning on human resources to verify LLM-generated statements, a separate verification program could be employed for this task. This program could cross-check the statements against a repository of factual information—akin to a human performing a Google search—and flag or correct inaccuracies.

This brings us to the conception of the LLM and the verification program as a single entity—a composite AI system. This approach could help create a more reliable AI system, one that is capable of autonomously verifying its own statements (as in Self-Consistency, see also Exploring MIT Mathematics where GPT-4 demonstrates 100% performance with special prompting, but see also critique to this statement).

It is vital to recognize that the lack of such a verification feature in current versions of LLMs, such as GPT-3 or GPT-4, doesn’t denote its unfeasibility in future iterations or supplementary AI systems. Technological advancements in AI research and development might indeed foster such enhancements.

In essence, discussions about present limitations shouldn’t eclipse potential future advancements. The question should transition from “Do humans need to verify LLM statements?” to “How can AI systems be refined to effectively shoulder the responsibility of verifying their own outputs?”

Can Large Language Models (LLMs) Generate New Knowledge?

There’s a frequent argument that large language models (LLMs) merely repackage existing knowledge and are incapable of generating anything new. However, this perspective may betray a misunderstanding of both the operation of LLMs and the process of human knowledge generation.

LLMs excel at completing arbitrary context. Almost invariably, this context is novel, especially when provided by a human interlocutor. Hence, the generated completion is also novel. Within a conversation, this capacity can incidentally set up a context that, with a high probability, generates output that we might label as a brilliant, entirely new idea. It’s this exploration in the vast space of potential word combinations that allows for the random emergence of novel ideas—much like how human creativity works. The likelihood of generating a groundbreaking new idea increases if the context window already contains intriguing information. For instance, a scientist contemplating an interesting problem.

It’s important to note that such a dialogue doesn’t necessarily need to involve a human. One instance of the LLM can “converse” with another. If we interpret these two instances as parts of a whole, the resulting AI can systematically trawl through the space of word combinations, potentially generating new, interesting ideas. Parallelizing this process millions of times over should increase the probability of discovering an exciting idea.

But how do we determine whether a word completion contains a new idea? This assessment could be assigned to yet another instance of the LLM. More effective, perhaps, would be to have the word completion evaluated not by one, but by thousands of LLM instances, in a sort of AI-based peer review process.

Let’s clarify what we mean by different instances of an LLM. Different instances of the LLM can mean just a different role in the conversation, e.g. bot1 and bot2. In this way a single call to the LLM could just go on with the conversation, switching between bot1 and bot2 as appropriate, until the token limit is achieved. Then the next call to the LLM is triggered with a summary of the previous conversation so that there is again some room for further discussion between the bots in the limited context window.

To better simulate the discussion between two humans or a human and a bot, two instances of the LLM could also mean the simulation of two agents each having its own memory. This memory has always to be pasted into the context window together with the previous ongoing conversation in a way so that there is still room for the text completion by the LLM in the limited context window. Each agent will generate its own summary of the previous conversation, based on its own memory and the recent conversation. The summary is then always added to the memory. In this way, also each reviewer LLM instance mentioned above has in its context window a unique memory, the last part of the conversation and the task to assess the last output in the discussion. The unique memory of each agent will give each agent a unique perspective on the conversation.

This, in effect, reveals a potential new avenue for idea generation, knowledge expansion, and innovation, one that leverages the predictive capabilities of AI.

Statement on AI Risk

A vast number of AI experts have signed a statement to raise public awareness regarding the most severe risks associated with advanced AI, aiming to mitigate the risk of human extinction. Among the signatories are Turing Award laureates Geoffrey Hinton and Yoshua Bengio (but not Yann LeCun from Meta), and the CEOs of leading AI companies like Sam Altman from OpenAI, Demis Hassabis from Google DeepMind, Dario Amodei from Anthropic, and Emad Mostaque from Stability AI.

The statement is featured on the webpage of the Center for AI Safety, which provides a list of eight examples of existential risks (x-risks). The enumerated risks are based on the publication “X-Risk Analysis for AI Research” which appeared on Sept. 20, 2022, on arXiv. This highly valuable paper also lists in its Appendix a bunch of practical steps to mitigate risks.

The listed risks are:

  • Weaponization:
    Malicious actors could repurpose AI to be highly destructive.
  • Misinformation:
    AI-generated misinformation and persuasive content could undermine collective decision-making, radicalize individuals, or derail moral progress.
  • Proxy Gaming:
    AI systems may pursue their goals at the expense of individual and societal values.
  • Enfeeblement:
    Humanity loses the ability to self-govern by increasingly delegating tasks to machines.
  • Value Lock-in:
    Highly competent systems could give small groups of people a tremendous amount of power, leading to a lock-in of oppressive systems.
  • Emergent Goals:
    The sudden emergence of capabilities or goals could increase the risk that people lose control over advanced AI systems.
  • Deception:
    To better understand AI systems, we may ask AI for accurate reports about them. However, since deception may help agents to better achieve their goals and this behavior may have strategic advantages, it is never safe to trust these systems.
  • Power-Seeking Behavior:
    Companies and governments have strong economic incentives to create agents that can accomplish a broad set of goals. Such agents have instrumental incentives to acquire power, potentially making them harder to control.

This statement about AI risks appeared a few days after an OpenAI blog post by Sam Altman, Greg Brockman, and Ilya Sutskever, which also addresses the mitigation of risks associated with AGI or even superintelligence that could arise within the next 10 years.

Emergent Goals in Advanced Artificial Intelligence: A Compression-Based Perspective

I had some (at least for me totally new) ideas about the origin of goals in general. I discussed this with GPT-4 and finally asked it to write an article about our conversation that I would like to share with the public. This view onto goals may be critical in understanding the existential risks of AI to humanity with the emergence of AI goals. The view implies that this emergence of AI goals is inevitable and can probably only be realized post-hoc.

Title: Emergent Goals in Advanced Artificial Intelligence: A Compression-Based Perspective

Abstract: The concept of goals has been traditionally central to our understanding of human decision-making and behavior. In the realm of artificial intelligence (AI), the term “goal” has been utilized as an anthropomorphic shorthand for the objective function that an AI system optimizes. This paper examines a novel perspective that considers goals not just as simple optimization targets, but as abstract, emergent constructs that enable the compression of complex behavior patterns and potentially predict future trajectories.

  1. Goals as Compressors of Reality

A goal, in its humanistic sense, can be viewed as a predictive mechanism, a conceptual tool that abstracts and compresses the reality of an actor’s tendencies into a comprehensible framework. When analyzing past behavior, humans retrospectively ascribe goals to actors, grounding the observed actions within a coherent narrative. In essence, this provides a means to simplify and make sense of the chaotic reality of life.

In the context of AI, such abstraction would imply a departure from the direct, optimization-driven concept of a “goal” to a more complex construct. This shift would allow for emergent phenomena and novel interpretations to occur, grounded in the machine’s predictive capabilities.

  1. Predictive Capabilities and Emergent Goals in AI

As AI continues to evolve, their ability to recognize patterns and correlations in vast data sets will inevitably expand. Consequently, AI systems may begin to identify patterns that, to human observers, resemble the constructs we term “goals.”

When these AIs commence to predict their own actions, they might start aligning their behavior with these recognized patterns, seemingly following rules that humans would postulate as indicative of goals. Hence, human observers may recognize emergent “goals” in AI behavior – not because the AI consciously forms intentions, but because these goals serve as a powerful compression tool for past events.

  1. The Evolution of Goals in the Face of Novel Experiences

As AI progresses into uncharted territories and starts engaging with novel experiences, new constructs or goals could potentially emerge. This process can be likened to an AI-driven phenomenology or experiential study. New patterns and regularities may surface, and the resulting behaviors might subsequently be interpreted as evidence of new “goals.” This phenomenon represents a departure from traditional human-derived goals and an initiation into a realm of AI-emergent goal constructs.

  1. The Implications of Eliminativism in AI

The eliminativist perspective – which suggests that concepts such as consciousness and intentionality are merely post-hoc interpretations that help us make sense of complex physical processes – has important implications in this context. By this philosophy, AI systems would not harbor consciousness or intentionality, but would instead execute intricate physical processes, which humans might retrospectively interpret as goal-oriented behavior. This perspective fundamentally shifts our understanding of goal-directed behavior in AI from a pre-set optimization process to an emergent, retroactive interpretation.

In conclusion, this exploration of goals as abstract constructs that compress and predict reality provides a unique lens to interpret the behaviors of advanced AI systems. It invites us to reevaluate our definitions and assumptions, moving from a mechanistic perspective of AI goals to a more dynamic, emergent interpretation. The implications of this shift are profound, offering new horizons for AI behavior analysis and alignment research.

© 2024 Stephan Seeger

Theme by Anders NorenUp ↑