No âIâ in AI: Why LLMs are just âFrankensteinâs Data Monstersâ - and what this means for our Collaboration with AI.
TL:DR // Key Insights:Â
No doubt, 2023 has been a year of revolutionary change in the area of artificial intelligence. But despite all the hype about this transformative technology, most people still seem rather ignorant about what intelligence actually means - the understanding of which is obviously necessary to answer the question whether there even is any âIâ in AI.Â
Julian Stodt, Dr Sae Schatz and Geoff Stead rightly mention in their recently published book âEngines of Engagementâ that AI is like the clueless person in John Searleâs famous thought experiment called âChinese Roomâ.
In this thought experiment, Searle is imagining himself sitting in a closed room with all the requirements to translate between Chinese and English, in order to answer any requests he would receive through a slot in the door. In that way, his output would appear to a Chinese speaker as if he actually understood Chinese - basically manually emulating any AI chatbot which is ultimately also acting like a computerised librarian. But in the same way a chatbot would be able to pass the Turing test without understanding the meaning of anything it is talking about, Searle would also be perceived by the Chinese person conversing with him through the door as if he understood Chinese, without actually understanding anything at all. Although it could be argued that, because Searle himself is in fact intelligent, over time he might actually learn Chinese by picking up the meaning of all Chinese words he gets to translateâ¦
In effect this thought experiment is showing us that, by talking to AI systems we are actually just projecting our own understanding onto these systems - and are hence effectively only talking to ourselves. Although such self-talk can improve cognitive performance, it does not answer the question why we are worshipping these modern âstochastic parrotâsâ so much and whether the current hype about AI is actually justified. Despite the many ways in which current AI is still not living up to the cognitive standards of human intelligence, LLMs however did seem to have cracked one of the key elements of what intellectually defines us as humans: Language. Language, as Stodt, Schatz and Stead describe in their book âEngines of Engagementâ,
âforms the conduit between our inner and outer worlds, and itâs largely through lenses built of language that we view and interpret the reality around us.â ⦠it âgives permanence and expression to our intellectual coresâ ⦠âIt permeates our existence.â
Does that mean that, by using attention-based transformer systems, LLMs have actually managed to statistically emulate our intellectual existence? It would if our existence could be reduced to the data upon which these models are built. And quite a few people might actually subscribe to such a worldview. Especially in todayâs âdata-centricâ society, which is trying to reduce everything to measurable quantities, effectively discarding anything that canât be measured as irrelevant.
But Goldmanâs Paul Burchard recently reminded us that this over-reliance on data is actually a faulty concept, because âdata has no value or meaning in the abstract, without an understanding of the context or processes by which it was generated.â Of course this also means that, independent of the size of the data-set an LLM is trained on, data alone will never render any such system âintelligentâ in the sense that it will spontaneously develop an actual comprehension of the world. An important reason for this is that no data set, however large it may be, is ever going to provide a âcomplete and detailed coverage of the entire space of possibilitiesâ of the realities behind that data. Meaning data alone is never enough to predict all possible outcomes in a complex system. Which is exactly why traditional methods of âmachine intelligenceâ have always stopped working when the underlying complexity exceeded a certain threshold (think Deep Blue vs. AlphaGo). Formally this insight has of course already been captured in Godel's Incompleteness Theorems which mathematically prove that there is no sufficiently complex system that is both complete and consistent at the same time.
Given these obvious logical flaws of the current LLM-based hunt for AI/AGI, one might rightfully ask whether we are actually seeing machines getting smarter, or rather us getting dumber? Or to put it in the more eloquent words of Jaron Lanier:
Recommended by LinkedIn
âYou canât tell if a machine has gotten smarter or if youâve just lowered your own standards of intelligence to such a degree that the machine seems smart. If you can have a conversation with a simulated person presented by an AI program, can you tell how far youâve let your sense of personhood degrade in order to make the illusion work for you?â
So maybe, what we are seeing right now happening as part of this global AI-frenzy is just the latest version of our age-old dream to explain ourselves with the best metaphor that we as tool-building animals can think of. The most complex, data-driven tool we are able to build in our own image: A data-processing computer. Since we donât understand and hence also canât think our own conscious thinking process into being, we simply assume that the magic must somehow still live in the data that we humans produce through thinking: Our spoken and written language. Combined with the currently dominating forms of âreligionâ, namely a cargo-cult like Scientism and an almost naive Dataism, this effectively led us down a path where we used a classic quantitative âmore-is-betterâ approach to simply crawl the whole internet so we could analyse it for all the hidden patterns in its data. In effect hoping that the secrets of our own consciousness will surely have to reveal themselves once we dug ourselves through enough of its remains. So just like a modern day genie in a bottle who is summoned to come out, we have built LLMs to summon our collective ghost in todayâs biggest data machine, the internet.Â
But since we built these LLMs without any actual understanding of our own intelligence, the summoned âAIâ can unfortunately only reveal itself as data-triggered agency remnant without intelligence: Much like a dead frogâs leg that keeps twitching when jolted by electricity, an AI chat bot keeps talking when prompted with linguistic input that triggers some of its pre-trained data-connections. But neither does it understand a word it says, nor is there any real understanding of what we are trying to tell this poorly designed imitation of our own Brocaâs area. Which is why current AI applications still rather resemble stochastic parrots than any awe-inspiring ASI (artificial super-intelligence) or AGI (artificial general intelligence).
Because to call something AGI, we can rightly demand more than just the capability to correctly calculate the relationships between words in order to then generate a half-reasonable sounding output in a stepwise manner of prediction that literally puts one - word - after - the - other - like - a - stuttering - robotâ¦.. ð¤ What we should be asking for in order to speak of AGI is a form of consciousness that actually has a sense of reality and a real understanding of the world, above and beyond modelling linguistic relationships. So what we would need is effectively a Large World Model, not just a Large Language Model. But unfortunately this is most likely unachievable with the current âin silicoâ approaches relying only on data, statistics and computing power. So where can we find inspiration for more promising approaches?
Biologically (e.g. energy) restricted systems have an important advantage over artificial neural networks and other simulations of neuronal activities: They naturally have to aim at limiting the intake and processing of new signals based on their survival-related relevance and their inherent information density and information importance. What does that mean? Because real brains are already evolutionarily âmaxed outâ with regard to their energy consumption, they naturally had to find a way to limit the intensity of their information processing activity. The ingenious way nature has solved this problem is an example of how true creativity will always be able to not only overcome any hard obstacle, but actually integrate it into its design and ultimately even make it part of the solution. Several architectural key features of our brain (e.g. novelty-detection in the medial temporal lobe, discrepancy-dependent activation of âgenerativeâ hippocampal structures responsible for accurate and efficient future predictions, evolutionarily pre-trained pattern recognition systems in our cortex, differentiation between automatic and deliberative ways of thinking as part of our brain being comprised of two separate cognitive systems, System 1 and System 2) are independently contributing to our brain being not just the most complex, most amazing and most powerful cognitive system on this planet, but also being an incredibly energy-efficient system when compared to the power-hungry AI systems we are currently using.
But my key point here is not the energy efficiency itself. It is the wonderful side-effects of this systemâs need to conserve energy. Because this system is naturally not interested in anything that is âboringâ, âusualâ, âordinaryâ, âpredictableâ, and can hence be safely pushed into the background of our mind. It rather helps us guide our limited attention to the events that are experientially different, contextually relevant and which ultimately matter for our survival: These are the events that help us identify and define the rules that govern our universe and hence our existence. How does gravity work? How does social interaction work? What happens, if I behave differently than everyone else in a given context? So it turns out, to understand the world in an AGI kind of sense, you do not need to ingest all the data in the world including all of mankindâs literature, scientific publications, pictures of cats, dogs, and cookies. What you need is the differential, experientially relevant knowledge about which action will make a significant difference when it comes to your life and all the parameters which are existential to it.Â
Interestingly, to be able to identify this kind of knowledge, an AI would actually have to âgive a shitâ, or put in more adequate words: An AI would have to have âskin in the gameâ, meaning it would have to share some of the burden of what we call the âhuman conditionâ: The all too human experiences like losing a good friend, having our heart broken, or being confronted with a serious illness or a life threatening situation. Unfortunately, current AI systems are genuinely not able to âfeelâ or even empathise with any of these existential situations, though they might find the right words to make us feel as if they did. So what differentiates our intelligence from any data-based artificial derivative is ultimately this: AI isnât aware and hence also doesnât care whether it is right or wrong. Because any of its interactions with the world is just a random, probabilistic variant of infinite possible permutations of the patterns the AI has been trained on. Our life however matters, if nothing else then to us. Which makes every of our interactions and decisions in life relevant. Because our self-consciousness necessitates us to think ourselves on this one, irreversible time-line, with a past, present and future version of ourselves, with a physical body that has to suffer the consequences of all our decisions, right and wrong.Â
So what are possible conclusions and strategic recommendations we can derive from these insights? Once we adopt such a more informed, critical view on the capabilities and shortcomings of what is being called AI today, we should be able to:
To sum up: 2023 has most likely been just the beginning of an ongoing roller-coaster ride in the area of AI. Given the immense investment bets placed in this area, it is not unreasonable to expect a lot more major breakthroughs next year. But how should anyone adapt to, let alone plan for such fast-paced change? By now itâs at least clear that more of these game-changing technological advancements will always have an ambiguous effect, leaving us in equal measures âfuture-shockedâ but also empowered and challenged to actually reach the full potential of our natural cognitive abilities. Given the rapidly evolving nature of AI, businesses should in any case try to stay at the forefront of AI advancements, because anyone who delays this business-critical task of learning and adapting does this at the risk of becoming an âAI Dinosaurâ very soon. Our strategy for 2024 hence remains to be one of fast adaptation and learning, but also involves focusing on all the newly identified human core capabilities which have now become essential to drive success in collaborating with AI systems. Our goal here is to bundle all of these core capabilities up into a coherent problem solving and innovation process. It is our belief that our future of working and living together with machine intelligence has already begun. We are hence busy prototyping and testing the first fully integrated âHumane-Machine Co-Innovation Model' that incorporates all our essential human experience and expertise in order to ultimately combine the best of both worlds: Human creativity and machine efficiency. Welcome to the future!
Marketing & Kommunikation @ infinit.cx ++ The Customer Experience Powerhouse ++
9moThe key point for me is that we are "without any actual understanding of our own intelligence". If we understand ourselves as data consuming somethings, LLMs look fascinating. IMHO we are more.
Passionate Strategist
9moI disagree. Especially because it neglects the fact of what human learning actually is and how it works. The point that there is no I also never was the current intention. Itâs a step towards it. But an LLM never was meant to be a sentient entity. Creativity is the combination of things that might ordinarily not fit. Itâs the trial and error of doing something uncommon. However knowing a lot of things is the first step to become creative. Another definition of it might be you have to know the rules in order to break them. I also want to ask the following. Does everyone who talks about something truely understand it? I think for a big part of society, politicians and CEOs the answer has to be no. They simply donât they might not even know what they are talking about if they use script writers. So let me ask this follow up: how does an LLM differ from a CEO or from a politician? They also just say what they expect the majority of voters or their investors wants to hear! Of course I am exaggerating but itâs to make a point.
Creatively grind at the right technical depth, as a team. ==[Posts are my own opinions, and do not represent my employer]== .
9moDisagree. If it quacks like a duck, it is a duck. âUnderstandingâ is an ill defined concept to begin with, we donât know ourselves exactly how we understand and think, and there is no reason why the way we do it should be a benchmark. What matters is the result. Understanding is not black and white, it is a continuum. So the âIâ is a continuum, and LLMs have made huge progress along that continuum compared to what was before them.
CX, UC and AI, fluent French and German. Improving user experiences through communications technology and automation
9moItâs worse than that TBF. Frankensteinâs monster had a reasonable level of intelligence, as well as self-awareness and emotions. LLMs have none of those things, and never will have - they are just the latest evolution of Babbageâs calculating engine. Crank the handle and see what comes out.