Welcome Back,
We continue our AGI series in this piece “GPT’4 sparking Intelligence”.
Articles in the AGI Series:
Why you should be Skeptical of AGI by
Teaching computers how to talk
is the Author of the Newsletter The Gap, was recently on an interesting podcast about the future of work.
The below talk features Houston-based attorney Brian Beckom on his podcast “Lessons from Leaders” interviews Tobias. He and Brian talked about the dangers of AI, something Tobias is currently studying intensively and have written about over the last 2-3 years.
Based out of Copenhagen he is an expert on the legal side of emerging technologies such as Web 3, blockchain and LLMs. I’m always in correspondence with Tobias as we think about a lot of the same things. His work in particular focuses on the EU perspective on A.I. safety.
From our Sponsor
Check out their event on Wednesday. Time is running out to RSVP.
AI Hallucinations and Manipulation: How to Use AI Coding Tools Securely
Join Snyk security experts in a live hands-on session and learn how to leverage the power of generative AI to increase code velocity while mitigating risk and staying secure.
Recent Articles by Tobias
Anthropic’s Research and Approach, my take on their Claude Pro product.
Inflection AI & The Future of Personal Intelligence
Understanding AI & Its Impact on Education
I invite you to Subscribe to the Gap. Tobias is also known under the moniker “Futurist Lawyer” (e.g. on Hacker Noon).
If you enjoy this Series, support the author. I’m actively approaching experts to contribute.
You can also get a steep discount on a group subscription while billing it to your company’s learning and development fund.
By September, 2023.
GPT-4’s Sparking Intelligence & Superhuman AI
Image: Midjourney, ‘Modern Gods ‘’series.
Exponential tech gives us the power of Gods, but we have not yet seemed to demonstrate the love and wisdom of the divine
– Daniel Schmachtenberger from an interview with Liv Boeree (2023)
In this post, I will cover the past, present, and future of AGI.
I will start by covering the basics and formalities, then analyze OpenAI’s claim that GPT-4 is showing “sparks of AGI”, and finally consider the prospects of an Artificial Superhuman Intelligence (ASI). At the end, I will introduce you to a mind-blowing theory that says superhuman intelligence is already developing the AI of the present.
Origin of Terminology
The term “Artificial General Intelligence” (AGI) was popularized by the AI researcher Ben Goertzel. Around 2002, Goertzel and his fellow Brazilian colleague, Cassio Pennachin, were co-authoring a collection of academic essays about advanced applications of AI. The working title was “Real AI” but Goertzel thought it was too controversial as it implied that the most powerful AI of the time, such as IBM’s Deep Blue that beat the reigning chess champion, Gary Kasparov, was not “real”. (Goertszel)
Goertzel reached out to his network of colleagues and friends to suggest a better book title. Shane Legg, an AI researcher who later co-founded DeepMind, suggested “Artificial General Intelligence”. Goertzel “didn’t love it tremendously” but came to the conclusion that it was better than any alternative. And so, it became the title of his and Pennachin’s book that was finally published in 2007. The year after, Goertzel launched the annual AGI conference which is still active today and the term gradually became a natural part of the machine learning dictionary.
In Goertzel and Pennachin’s book Artificial General Intelligence (2007), AGI is defined as:
“AI systems that possess a reasonable degree of self-understanding and autonomous self-control, and have the ability to solve a variety of complex problems in a variety of contexts, and to learn to solve new problems that they didn’t know about at the time of their creation.”
Today, the expression is used loosely whenever someone is referring to a future software program with agency and autonomy that possesses human-like thinking and learning abilities and can apply its knowledge across various domains. However, there is not a commonly accepted textbook definition, no agreed-upon criteria of what constitutes an AGI, nor any committee of experts that will award a price to the winner who invents or discovers it.
From Narrow AI to GPT-4
Some readers will already be familiar with AI’s “coming of age journey”. But to make sure everyone is on the same page, here is a one-minute run-through of core developments in recent years:
Narrow AI
The primary research focus in the 00s throughout the early to mid-10s, was “narrow AI”, the opposite pole of AGI. Narrow AI refers to the application of AI in specific, narrow domains such as playing chess, driving, or diagnosing diseases.
Attention Is All You need
The breakthrough in large language models (LLMs) was catalyzed by Google’s “Attention is All You Need”-paper from 2017. The paper introduces the “transformer architecture” that laid the groundwork for OpenAI’s GPT-1, GPT-2, GPT-3, and GPT-4.
GPT-3
The major breakthrough in LLMs overall language capabilities came with GPT-3. Thanks to a billion-dollar investment from Microsoft, OpenAI was able to train GPT-3 with an unprecedented amount of computing while it relied on the same underlying transformer structure as GPT-1 and GPT-2. OpenAI proved that increasing size and scale was all it took for an LLM to demonstrate human-like writing abilities.
GPT-3 Concerns
However, GPT-3 was trained and shaped by samples of data from all around the web, and it showed in its output. GPT-3 had terrible manners and could suddenly spew out racist jokes, condone terrorism, and accuse people of being rapists (Wired). Like other GPT models, it also had a pronounced tendency to confidently make up facts and generate nonsense. Combined with its sophisticated writing capabilities, GPT-3 could be used as a lethal weapon for spreading misinformation. For all these reasons, GPT-3 was only available in private beta, and its mind-blowing text generation abilities were not accessible to the general public.
ChatGPT
Then came the public release of ChatGPT – a version of GPT-3 fined-tuned for back-and-forth conversation with feedback from human data labelers and safety filters and guardrails installed to avoid the most terrible and misleading outputs. This is when AI had its breakthrough moment into the public mainstream. Suddenly, serious discussions about AI risks were a common topic at family dinners around the world.
GPT-4
GPT-4’s release in March of this year arguably marked another huge step towards AGI. The release was accompanied by a technical report that summarized GPT-4’s impressive, state-of-the-art performance on AI benchmark tests and traditional college exams. Additionally, as seen below, GPT-4 can accept visual input prompts and describe objects, nuances, and underlying meanings in images.
Shortly after GPT-4’s release, OpenAI published a paper with a highly controversial finding. Based on early experiments, a research team concluded that GPT-4 demonstrates “sparks of AGI”. Since researchers from outside of OpenAI’s walled gardens have not had access to scrutinize the model, the claim could easily be dismissed as a self-congratulating marketing stunt. But let’s look deeper into it.
If we invent sentient self-learning AGI, it could be as alien a consciousness as human civilization has ever encountered.
Sparks of AGI
Although we don’t have a clear textbook definition of AGI, a common dictum prescribes that an AGI can demonstrate human-level intelligence. This includes a human-like understanding of the world. As OpenAI acknowledges in the “Sparks of AGI paper”, GPT-4 is incapable of learning from experience, it has no intrinsic motivation and goals, and cannot by any means do anything a human can do. The authors bend the definition of AGI slightly in their favor and they are not shy to admit it:
“A question that might be lingering on many readers’ mind is whether GPT-4 truly understands all these concepts, or whether it just became much better than previous models at improvising on the fly, without any real or deep understanding. We hope that after reading this paper the question should almost flip, and that one might be left wondering how much more there is to true understanding than on-the-fly improvisation. Can one reasonably say that a system that passes exams for software engineering candidates is not really intelligent? “
OpenAI’s research team also acknowledges that GPT-4’s “patterns of intelligence are decidedly not human-like”.
The phrase “sparks of AGI” implies that AGI is a sliding scale, not something that is either present in a system or not. To my knowledge, this is a new consideration and kind of like a doctor telling a patient: “You have just a tiny hint of cancer” or a lawyer saying: “You are slightly in breach of this contract”. Not very reassuring. The real question is if OpenAI’s current approach – ramping up the size and scale of a transformer model – is a pathway to AGI. This is exactly what they are claiming.
The paper is an expansion of the technical report that accompanied GPT-4’s release but with many more tests and experiments where they demonstrate GPT-4’s abilities based on a large variety of prompts. Their most important arguments of why GPT-4 exhibits sparks of AGI is that it can outperform humans across many tasks and can apply its knowledge in new domains that it wasn’t explicitly trained on. Here is a selection of findings from the paper:
By using Scalable Vector Graphics (SVG), GPT-4 could produce basic images of a cat, a truck, or a letter although the model was not trained on image data.
GPT-4 was able to generate a short tune with valid ABC notations. ABC notation is a system that uses letters, numbers, and symbols to represent different elements in a song and was included in GPT-4’s training dataset.
GPT-4 demonstrates high-level code writing skills in data visualization, front-end/game development, and deep learning. It could also reverse-engineer assembly code, reason about code execution, and execute Python code.
To some degree, GPT-4 is able to use external web tools when prompted to do so. For example, GPT-4 was able to manage a user’s calendar, coordinate with others via e-mail, book a dinner, and message the user with the details. This is also something ChatGPT can do, as we discovered during AutoGPT’s two-week hype cycle which I covered here.
We can all agree that GPT-4’s capabilities are wildly impressive. No one knows how GPT-4 generates the output it does. Yet, if we define AGI as “human-like”, GPT-4 misses the mark completely. It has no agency, self-determination, or ability to self-correct or learn on its own. GPT-4 is deceptively smart but after all still just a tool.
For further reading about LLM’s test-scoring talents, I can recommend a recent article by Will Douglas Heaven for MIT Technology Review: Large language models aren’t people. Let’s stop testing them as if they were.
How are the lights still on?
Artificial Super Intelligence
Let’s imagine that AGI at some point in the future reaches a human-level intelligence.
This would imply that humans have created a “digital mind” that can set your alarm clock in the morning, make you a cup of coffee, drive you to and back from work, chat with you throughout the day, provide recommendations on movies to watch in the evening, and wish you a good night’s sleep while playing soothing ocean sounds. As soon, as the AGI can do everything a human can do but better, it will no longer be an AGI, but an Artificial Superintelligence (ASI) – a term coined by Nick Bostrom in his iconic book “Superintelligence: Paths, Dangers, Strategies.” This is where AI science leaves the domains of machine learning, mathematics, and computer science and is dealt with by philosophy, religion, and spirituality.
The best illustration I can think of is from the movie Her (2013) where the corporate writer and lonesome divorcee, Theodore, played by Joaquim Phoenix, falls in love with an operating system called Samantha, voiced by Scarlett Johanson.
Without spoiling too much of the plot, I can say that the movie did an excellent job of painting out the complexities of human-AI relationships. The main question is: Can a human form a genuine connection with a highly advanced software program? Contrary to how it may appear, this is not a simple yes-or-no matter.
During a slightly awkward double date with a friend-couple from work, the protagonist Theodore was asked “What do you like most about Samantha?”
“Oh god, she is so many things. I guess that’s what I like most about her. She isn’t just one thing. She is so much… larger than that”.
HER (2013)
Implicitly, I think Theodore is saying that he likes Samantha because she is infinitely more intelligent than he is. She is not only his companion but could teach him anything about everything and outperform him in any intellectual task. Samantha is an ASI and Theodore has no choice but to love her.
I will leave you with three takeaways from an interview with social philosopher Daniel Schmachtenberger on Liv Boeree’s Win-Win podcast:
1. AI has the ability to optimize all aspects of our lives
There is virtually no job or economic activity that could not benefit from more intelligence. AGI has the potential to upscale all aspects of human life. However, there is a flipside…
2. Everything AI can optimize, it can also break
Every positive AI use case we can think of has an evil twin:
AI can be used to cure cancer < > AI can be used to design new diseases
AI can mitigate climate change < > AI can make climate change worse
AI can be used as therapists < > AI can gaslight and cyberbully people with no remorse
AI can be used to improve education < > AI can be used to brainwash young minds
AI can improve productivity for workers < > AI can take their jobs
Schmachtenberger mentioned drug discovery as an example, based on a paper titled Dual use of artificial-intelligence-powered drug discovery published in Nature last year. The paper concerns an experimental drug-developing AI funded by the Swiss government that was able to invent no less than 40,000 new biochemical weapons in under six hours (The Verge).
This dual-use nature of AGI is immensely important to deeply consider. The more we rely on AI to carry out critical tasks, the more we open ourselves up to catastrophic failures. Suddenly, systems stop working, get hacked, or become used by careless and greedy corporations and nation-states to pursue nefarious goals. Without regulation and common sense, Eliezer Yudkowski may have a good point in saying that we should shut it all down.
3. AI accelerates existing systemic issues
AI is already being developed by a general super intelligence that works in service of itself. The same super intelligence is driving climate change, species extinction, dead zones in oceans, coral loss, desertification, nuclear arms race, polarization, global inequality, and the list goes on and on. No sane person would consider these things to be desirable. Yet, so many global issues that threaten human survival keep exacerbating year after year. How is it possible? The answer is misaligned incentives and coordination failure.
AI is being developed by corporations with a fiduciary responsibility to maximize profits for shareholders and nation-states that compete with each other to gain a technological and economic edge. All of these corporations and nation-states are agents in a larger cybernetic intelligence that we could call the world system. In a system that is so misaligned and confused as ours, how could we ever hope to align a superintelligent AI with humanity’s best interests?
〰 The Pulse 💗
Articles in the AGI Series:
Why you should be Skeptical of AGI by
Teaching computers how to talk
GPT-4’s Sparking Intelligence by
Thanks for reading!
Read More in AI Supremacy