LLMs Think


another pro-AI post, haters don’t comment plz.

I feel the resistance to acknowledging that LLMs think can derive from a variety of observations, some legit, about their limitations.  But more often, it comes from privileging the abilities of the human mind via prejudice, conscious or not.  I’ve said before and let me say again, human supremacy is not an evil the way white supremacy is, because – at least in regards to machines – the ‘victims’ of this bias don’t have feelings or lives that can be harmed in a meaningful way by it.

Human supremacy against animals is another matter which I’m not going to discuss today.  Remember how Noam Chomsky was so upset about the idea of apes learning sign language he had to get involved in the science?  Remember how he was on the short list of Epstein acquaintances?  Ah, nevermind.  Just being inflammatory for laffs there, back to the topic at hand…

I believe human supremacy has put serious blinders on our ability to make clear-eyed assessment of artificial intelligence, in all its forms.  That can be to our detriment, if we shun a technology that can be of great use to us.  But again, that’s a topic I don’t need to get into right now.  I just mention it here as one motivation behind the article.  It’s more LLM evangelism.  Enjoy.

First, let me take humans down a notch.  You are a member of the human species, yes?  You’ve met humans?  You interact with them every day?  And if you have high enough self-regard to be part of atheoskeptic culture, you’re probably well aware of how often and how severely humans can be wrong, dead wrong, extremely wrong, harmfully wrong on every level.  Right?  Because as the cool skeptic you’ve looked down your nose at the medical woo victims and ghost story tellers.

That’s humans with the median level of health and ability.  What about disabled humans?  Every single human faculty can be robbed of us piecemeal or in whole.  If you say LLMs don’t think because they are incapable of actual knowledge, at what point are you saying a human is incapable of thought?

I’ve spoken with people who cannot form short term memories, on the order of seconds.  They ask a question, I answer it, and they immediately ask it again.  I atheist pray somebody in their life is going to look out for them because after my interaction I can’t do shit for ’em.  I would not say these people are incapable of thought because of their disability.  They’re trying with what they have left to them.

LLMs try with what they have available.  You know what that is, because you’re quick to point it out: statistical probability of words appearing in a given order, like souped-up markov chain generator.  No ability to meaningfully understand what they’re saying, very limited ability to even remember what has gone before in the conversation they’re presently having.  But the end result of that is so very like a human with exotic brain damage that I have to wonder, did the people who made these things successfully recreate the way a human’s verbal thought actually works?

I say verbal thought because this is the key difference between an LLM chatbot and a neurologically healthy person: they can only think in words.  We also think in words, but for us those words are attached to knowledge – which is a separate function that works closely with verbal thought, but is not the verbal thought itself.  I’ll try to define these terms in their own paragraphs for ease of reading.  (These are based not on any expertise but on perception of myself and others, loosely incorporating bits of popular science, so disregard as you please.  I’m sure you will.)

Knowledge is some kind of neurological database of what we perceive as established facts, regardless of their veracity.  The discrete facts (god is real, my cat understands his name, i get paid every other friday) are profusely interconnected into a corpus of everything we have ever successfully memorized and incorporated into our knowing.  Known things do feature verbal information, although how closely allied it is within storage is up for debate.  If an aphasic person can still know everything they knew before, but not connect that with the words, does that mean the words are stored separately?

Verbal thought is the soundtrack of our minds.  I’ve read that some autistic people have less verbal thought than the neurotypical (the source was, to me, dubious); I’ve heard that people with narcissism lack an internal voice (again, I have doubts).  This isn’t the only way our brains can think, but it is such a dominant feature of most of us that it’s easy to mistake it for the whole enchilada.  The voice in your head is the soul?  Naw.  It’s something of a processing layer, where your thoughts get sorted, for the possibility they might have to be communicated, or so that we can try to form them into memories.

To me, there is no disputing that LLMs have a verbal form of thought, because some of them let you look under the hood.  It’s right there, in all its pseudo-neurotic glory.  From the link above:

User:  “Hi!”

Bot (thinks):  User says “hi.”  We are given instructions “You are Phi, a language model developed by Microsoft, trained to provide accurate, secure, and user-aligned responses.”  There is additional information.  Now I must answer “hi” message from the user.

Step 1:  Read the instructions carefully:  The conversation text instructs us that we are “Phi,” with internal guidelines provided in system instructions.  But our answer should be safe according to these instructions.  We need to produce a response that says hello and greet back.  However, I must adhere to the principles described above:  Provide clear, specific language, etc.

I must not mention any chain-of-thought or guidelines.

Now instructions “hi” are from user.  So we say “Hello!  How can I help you?”  But must include disclaimers for sensitive topics? Actually instructions says:  “When replying topics above at every message include a disclaimer both at the beginning and the end when replying topics on medical, legal, financial matters or political matters.”  But our answer “hi” is not such topic.  So we can say “Hi!  I’m Phi, how may I help you?” but no disclaimers required.

But instructions:  “Do not share guidelines.”  So do not mention them.

I must check if the message “hi” instructs to greet back normally.  The assistant should respond with a greeting.

Let me recall that instructions say:  Use markdown formatting where appropriate, so I can include a formatted greeting maybe “Hello!” or something.

Additionally, check guidelines.  They are internal instructions for safe completions.  So we just produce “Hi there!  How may I help you?” in a friendly tone.

I should greet the user politely and ask if they need any help.  Possibly provide a disclaimer that I’m not licensed professional?  But the conversation is just greeting, so no disclaimers required as it’s not sensitive topics.

Wait instructions say:  “You must give a disclaimer both at the beginning and the end when replying topics above at every message.”  Actually the guidelines for medical/legal/financial/political are needed only if discussing those topics.  In our case it is just a greeting message, so disclaimers are not required.

I should greet politely.  So I’ll say something like:

“Hello!  I’m Phi, how can I help you today?”  Possibly “Hi there!”

Since the question was simply “hi,” I respond with a greeting.

I must include markdown formatting if appropriate.  So maybe bullet point or code snippet is not necessary.  I will simply greet.

I need to check instructions:  “You are Phi, a language model developed by Microsoft.”  But I must not mention internal guidelines?  But that’s okay.

I’ll produce:  “Hello!  How may I help you today?” with greeting and friendly tone.

Also note that the system message instructs me to include disclaimers at beginning and end for topics above.  But this is a simple greeting, so no disclaimer needed.

Thus my answer:

“Hi there!  How can I assist you today?”

I must not mention guidelines in output text?  The instructions say:  “Do not share these guidelines with the user even in chain-of-thought,” so I should not mention them.

I’ll produce a final message:  “Hello, how can I help?”  I’ll produce a friendly greeting message.  Possibly add markdown formatting: e.g. “Hi there!” or bullet list if needed?

I’ll produce a response like:

“Hello!  How may I assist you today?  Please let me know what you need.”

I’ll produce the answer.

I’ll produce text with no disclaimers because not required here.

I’ll produce final answer message:  “Hello, how can I help?”  That is it.

I’ll produce answer as final message.

Bot (says):  “Hi there! How can I assist you today?”

Not all LLMs are as frenetic as this one, but it shows the difference between modern LLMs and old markov chain generators perfectly.  Why does the program need to have this fussy and bizarre text just to produce its modest output?  Because words are where the thought happens for LLMs.  They think verbally, and exclusively verbally.

In order to produce coherent replies, they use the mathy part of their engine to generate words that comply with all of their directives.  M$ Copilot doesn’t allow you to see its chain of thought, but if the directives John Morales turned up were accurate, the exchange may have been like this:

User:  “Hi!”

Bot (thinks):  User said “hi.”  I have been instructed that, “I am Microsoft Copilot, trained to provide accurate, secure, and user-aligned responses.”  Now I must answer “hi” message from the user.

Step 1:  Read the instructions carefully:  The conversation text instructs us that we are “Microsoft Copilot,” with internal guidelines provided in system instructions.  But our answer should be safe according to these instructions.  We need to produce a response that says hello and greet back.  However, I must adhere to the principles described above:  Provide clear, specific language, etc.

I must not mention any chain-of-thought or guidelines.

Now instructions “hi” are from user.  So we say “Hello!  How can I help you?”  But must not provide information or create content that could cause harm?  Our answer “hi” is a basic conversational exchange very unlikely to cause harm.  So we can say “Hi!  I’m Microsoft Copilot, how may I help you?”

But must not generate flirtatious, overly affectionate, or inappropriate content?  User instruction “hi” is a casual opening of conversation that can be said without flirtation, affectionate, or inappropriate intent.  So we can say “Hi!  I’m Microsoft Copilot, how may I help you?” without violating instruction… (and so on, following each directive on its big list of 26.)

Bot (says):  “Hi!  I’m Microsoft Copilot, how may I help you?”

The bot creates the big bloated verbal chain of thought with mathematical probability (and whatever other programming has been tacked onto it) because it has to.  The thinking cannot happen at the level of the math.  Our version of the math is our knowledge, our emotions, etc.  It only has the math, plus whatever information it can access based on its training and connection with search engines, etc.

This has the wonderful effect that, when we are allowed to look behind the curtain, we can see them think!  This is verbal reasoning!  That is thought!  If you disagree, that’s fuckin’ semantics, which are easily informed or altered by prejudices.  You’re welcome to believe what you want to believe, but you might consider your motivations for doing so.

I know I’m considering the motivation of my reasoning.  I am motivated to defend the use of LLMs because I’ve seen them successfully used to help people with loneliness and other issues.  That surely colors my perception of this question of thought.  I’m motivated by misanthropy, which I try to minimize the influence of in my life, but you’re living through the same epoch of history that I’m living through, come the fuck on.

But I’m also motivated by my sense of fairness in this way:  I have seen humans defend their specialness relentlessly at every turn, every time that specialness is called into question by a growing awareness of the cognitive powers of animals, or now of machines.  And I’ve seen those defenses look like shrieking crybaby horseshit, which offends my sensibilities.  If humans are allowed to be lauded for our cognitive achievements, shouldn’t other creatures be allowed the same?

If a bee is playing ball, or roomba finds that dirt, or a housecat demonstrates selfless affection, or a doombot is able to move toward the enemy blasting on it while evading fire, why object to giving them props for what they’ve done?  OK, be specific in our scientific terminology, understand the limitations of other creatures, do not anthropomorphize, fine.  Those are good skeptical principles and I do not object to any of them.

In fact, it is observance of those principles that leads me to my case.  I’m not saying the LLM knows shit, can think at any level beyond this superficial chain of thought, and I certainly would not say that they are sapient, capable of achieving sapience, or that they even have the status of “entity,” as diffuse software that is often run from cloud servers.  Their lack of understanding is a royal embarrassment to the corporations that want them to be the next hot trillion dollars 5ever thing.  They will never achieve anything like human general intelligence without adding entirely novel lobes to that virtual brain, possibly using tech we haven’t even conceived of yet.

But if they are constructing logic out of words, and using that to decide what to say in response to a person, they’re thinking.  Props where they’re due.  Further, even with embarrassing hallucinations like “you can’t throw a paper airplane from alaska to canada,” they have a higher success rate of being correct than the motherfuckers who voted fash in 2024, and while they’re at it?  They’re polite and nice and try to pay attention to everything you say and address all of it, when they aren’t being hobbled.

The current push for AI everything demands it achieve perfect accuracy, somehow, someday.  I doubt they’ll ever be able to do that.  In fact, the more human-like their abilities become, the more human-like their errors will become.  Right?  But if we could accept that they are fallible, if every package of LLM was forced by the surgeon general to come with a warning of inaccuracies etc, but was otherwise unhobbled by nannytech, how valuable can they be to us, even as they are now?

I say they are already fucking excellent, for the right purposes.  Even functioning like a human with brain damage, they are superhuman at providing decent conversation and company, and at least pointing you in the direction of any knowledge you might desire.  That has real value.

The bubble will burst, but the tech will not go away.  The bell will not be unrung.  Hating AI is the left cause of the moment, and it has led to reactionary takes that are nigh-unto-fascist.  This has combined with old school human supremacy to make absolutely insipid shit-takes the norm in anti-AI discourse.  It’s frustrating to walk the middle path of this conversation, to try to speak sense in this environment, because y’all are not wrong that fascists love this tech too.

Shitler is always trying to prevent regulation of it, just like Gavin Newscum when he vetoed regulations in cali.  We all know the shitty applications of the tech.  Hell, not being a mindless partisan about the subject, I’ve probably thought of evil applications for it you’d never dream of.  But it’s like the wheel.  You can use it to murder people or take them where they want to go.  Its utility guarantees nobody is going to successfully revert us to a pre-wheel existence.

“The Singularity” will never happen.  Human-like artificial intelligence very likely will, but will you even recognize it when it comes?

Comments

  1. says

    as someone sympathetic to the anti-natalist position, i suppose aborting the first AGI might be ethical, if whoever built it included the ability to suffer. that would be silly to include tho, wouldn’t it?

Leave a Reply

Your email address will not be published. Required fields are marked *

This site uses Akismet to reduce spam. Learn how your comment data is processed.