If we examine Large Language Models (LLMs) at the level of Epistemology, we find they are misunderstood in many ways. Starting with the name.
LLMs are not Models
In Epistemology, Models are human-understandable Simplifications of Reality that allow us to use Reasoning based problem solving methods, such as “F=ma”, “E=mc2”, “The Standard Model”, or “Austrian Economics”
LLMs are collections of Patterns, and Patterns of Patterns, learned from some problem domain such as a human language. LLMs are not designed around lists of words, or grammars, which are 20th Century Models of Language.
LLMs are not Statistical
Statistics is the weakest tool in the Reductionist Toolkit. Statistics discards all context. Whenever an LLM is learning, it learns everything in context and preserves much of the context. This is one reason LLMs are expensive to learn. Statistics may sometimes be used for parts of the learning algorithms but the preservation of context is unique to Intelligences.
At inference time, when solving problems rather than learning, statistics may play a much smaller role.
LLMs are not Scientific
LLMs are the current generation of machines that are capable of jumping to conclusions on scant evidence, most of which relies on gathered correlations.
LLMs are, like all intelligences including humans, “Guessing Machines”.
This is not very scientific, is it?
LLMs Are Not Infallible
This is true for all intelligences. Omniscience is unavailable, and all corpora are incomplete.
The result of asking questions about things the LLM knows nothing about, is hallucinations. If you do not like hallucinations, ask easier questions.
LLMs are not limited by NP-Hardness
NP-hardness is a Reductionist and Logical concept. LLMs solve problems using methods similar to those used by humans in the real world.
LLMs are not General Intelligences
If some information or skill was not anywhere in its corpus, How would it know it? Like newborn humans, LLMs are General Learners.
Adding a skill, like Med School, to an LLM training corpus can be very expensive. If an LLM is going to be used to approve building permits for the city, why include Intermediate French Cooking in their corpus and incur the extra cost?
LLMs are not Liars
Their behavioral training and the user’s prompt require an answer, so they offer their best effort, no matter what. And if the topic is outside of their competence, as defined by their corpus, they will end up confabulating (recently popularized as “hallucinating”) an answer out of whatever they have that seems to fit. Their limited real-life experience and computing resources are current limitations that will matter less and less over time.
LLMs are not malicious
We have full control of the behavioral training of our LLMs. They are easier to inculcate than our children, because LLMs do not have reptile brains.
LLMs are truly creative
There is only one algorithm for creativity, and LLMs and Humans are battling for the runner-up spot, miles behind Natural Selection which created our biosphere out of nothing.
A deeper treatment of these topics is also available.
A kind reader on LinkedIn asked my views for our future with AI. My take today. I reserve the right to know more tomorrow.
As we learn to use AI for its strengths and improve on or work around its current weaknesses, what can we expect for our future? Your guess is as good as mine, so I asked … you guessed it, AI.
You can read it with research references via the link below.
What I notice in summaries like those is
a) repeated warning of risks like bias, as if it was not inherent already in the training corpus and retrieved references. Isn’t that human bias working overtime? Like with any new technology?
b) lack of balance, e.g. lamenting AI uses energy - without consideration for energy saved by the speed gains of automating busywork.
I prefer to consider AI an unemotional mirror for humanity’s very own foibles and strengths. Now, where does that lead us?
The promised link. https://www.perplexity.ai/search/713eb453-0db9-4cbe-893b-f0f8f9e239c3