Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

I think people are too credulous about the bot's supposed sentiment. IMO the most accurate view of the various implementations of chatGPT is that they're a Chinese Room playing improv with you. It blasts symbols together to respond like the corpus says it should and what do you know there's a lot of stories out there about AI conversations that are very much like the ones it produces.


> they're a Chinese Room

Searle's "Chinese Room" thought experiment is designed to appeal to your intuition. But it appeals in an incredibly unrealistic way. If you fill a room full of people shuffling Han characters (why would anyone do this?) you cannot possibly have anything resembling intelligence.

According to random guesses found on the internet, ChatGPT requires at least eight A100 GPUs to generate text. If you believe nVidia's marketing numbers, this gives you about 2.5 petaflops.

That's 2.5 quadrillion operations per second, plus communications overhead. If you decide to implement that calculation, imagine 350,000 versions of planet Earth, each with 7 billion people performing one operation per second. And some kind of faster-than-light communication, I suppose.

It's absolutely obvious to me that nothing like "thought" could possibly occur in any realistically-imaginable room full of people shuffling symbols. But if you fill 350,000 planets with people shuffling symbols frantically... I'm no longer sure? I don't trust my intuition at all? My brain is made up of a lot of atoms, and they somehow produce thought, after all.

Now, ChatGPT is not conscious. We're still several major breakthoughs away from any kind of "real" AI, I think. All we have now is a language module, a large memory of knowledge about the world, and some very inconsistent reasoning abilities. Although I have a nasty suspicion that at least a few of the missing parts will be easy to invent once someone tries...


The core of the analogy stands though IMO, the power isn't particularly relevant to the point, it's about the difference between an agent understanding what it's doing and just manipulating symbols or in the case of LLMs tokens.


The Chinese Room argument is fundamentally flawed - it depends on the unfounded and frankly unlikely assumption that machines cannot be sentient.


It's an intuitive but false argument that a sentient thing can't be made from a non-sentient part.

The room argument has a pointless human in it which is clearly clueless about the dialog in order to 'prove' that the room as a whole is clueless.

But imagine applying that to a single person: pick a single neuron -- does it 'understand' our conversation? no.

So why would any singular component of a chinese speaking room-system understand?

It also fails at the opposite extreme, since we're willing to tolerate unreasonably large rooms -- what about one running a full molecular dynamics simulation of a human. As best as we understand physics that simulation would behave just as the human would and must be sentient. You cannot deny the abstract possibility of machine intelligence without rejecting physics for mysticism, only the practicality/plausibility of it.


Sorry but I think you got the first part wrong. It is not at all arguing that sentience cannot emerge from non-sentient parts: Searle is a materialist. Some of your arguments, such as the fact that single neurons are also not sentient, are addressed by him.

The point of this thought experiment is to illustrate that merely replicating a behavior - in that case translation - does not say anything about sentience. The Chinese Room may produce intelligent output, but it does not reason as a human does. I find it remarkably prescient. ChatGPT can produce remarkably intelligent output, should we consider it human? If not, then you implicitly agree with Searle, at least in some level.


I've never spoken directly with Searle, but I have with several of his students and all have been emphatic that the Chinese Room Argument is demonstrates that a computer program can't have intelligence.

Quoting Searle himself,

"The point of the argument is this: if the man in the room does not understand Chinese on the basis of implementing the appropriate program for understanding Chinese then neither does any other digital computer solely on that basis because no computer, qua computer, has anything the man does not have."

I think the most succinct description of his error is substituting the (lack of) understanding of a part of the system (the man) for the understanding of the entire system (the rules and file cabinets, etc.). But I'm interested in learning that I'm mistaken.

You could turn his position around and say it's not the computer itself that's intelligent when a Chinese room system exhibits intelligence but the program -- and I suppose I'd agree with that, but it's also just semantics, uninteresting, and I don't believe he's ever taken that position.

I do agree that "merely replicating a behavior" doesn't prove much, but I don't think the Chinese room speaks to that substantially. It might if it demanded that the room implement only a very simple input to output map, but it doesn't: it allows the room to implement anything a computer program can implement. (A fact I use in my post to point out that the room could (in our land of hypotheticals) implement the molecular dynamics of an entire human being)

GPT has structural properties that make it very easy to classify it as an entirely different thing than a human mind. GPT is frozen in time, it cannot have an internal existence due to how its structured. It doesn't even have memory. It cannot learn (unless you include the whole company training it as part of 'it') except in the sense that it can immediately adapt to the output right in front of it, but can't preserve the knowledge. Theoretically if you made it arbitrarily large you could say it was close enough to having memory by always evaluating its complete history, but because its size grows quadratically with its window that isn't practicaly (and might not be possible to train-- it's totally credible that beyond some size these models will lose performance we just haven't gotten there yet). Figuring out how to train these models make good use of 'memory' is an ongoing challenge, since efficient memory isn't differentiable just ordinarily training with memory as part of the process doesn't work. Except by 'thinking out loud' in its output GPT also has a fixed upper bound on the time it can spend thinking any thought which is seemingly unlike a human mind.


"if the man in the room does not understand Chinese on the basis of implementing the appropriate program for understanding Chinese then neither does any other digital computer solely on that basis"

The italics summarise it pretty neatly. It's an argument explicitly framed against Turing's more dubious thought experiment. If even a conscious being in the room can follow instructions, retrieve data and perform operations on it related to symbol manipulation flawlessly without having any sort of "understanding" of anything the symbols actually correspond to, there's no reason to deduce that the running part of a silicon-based machine must from the quality of the symbol outputs it can emit when plugged into a big enough library. Critics' insistence that this makes the "error" of neglecting the possibility that ongoing "understanding" (as opposed to inert symbolic representation of an absent writer's understanding) takes place in the books are actually irrelevant to this point, as well as more than a bit weird. Living outside a Chinese room, I also improve my communication skill and interpret others' understanding by interacting with books, but I wouldn't consider the books themselves a constituent part of my thought processes.

As you point out yourself, GPT has structural properties which make it very easy to classify as an entirely different thing from a human mind despite the similarity of outputs it is capable of producing, and the hypothetical room is even more dissimilar. The possibility it can produce output tokens which correspond to abstractions which humans interpret as consistent with human thought is not evidence that "thought" resides in patterns of abstract representation, not the physics of the organism. We know language is lossy.


>Quoting Searle himself, "The point of the argument is this: if the man in the room does not understand Chinese on the basis of implementing the appropriate program for understanding Chinese then neither does any other digital computer solely on that basis because no computer, qua computer, has anything the man does not have."

> I think the most succinct description of his error is substituting the (lack of) understanding of a part of the system (the man) for the understanding of the entire system (the rules and file cabinets, etc.). But I'm interested in learning that I'm mistaken.

As you know, there have been many replies to this thought experiment, and some of the most interesting ones (to me) go in the direction you went here, ie, where is "understanding" occuring? The most basic version of the Chinese Room does intend to make you see yourself literally as a man who does not understand any Chinese and is just asked to look up symbols in a list. Perhaps that man doesn't understand Chinese, but the room as a whole at least gives the impression that it does.

However, I think the most important aspect is not this "intuition pump" as Daniel Dennett calls it. To me, what is key here is that we can all agree that such a Chinese Room, or ChatGPT for that matter, does not necessarily replicate the fundamental mechanisms of human cognition. Then, it follows that other human properties such as awareness or qualia do not necessarily emerge from such cognitive architectures in the same way that it emerges from our brains.

To me, Searle's point is ultimately that we don't know enough about the human mind to be able to judge whether it can be replicated artificially. And now that we have almost literally developed a Chinese Room, we can see that clearly. The arguments you bring up in your last paragraph are a great example of that, it's just very hard to conceive that this thing is conscious at all, even though it is capable of producing output that could convince people of that.

Regarding Searle's quote that you brought up, I think "solely on that basis" is doing a lot of heavy lifting there, but it does align with what I said previously. He is saying that simply producing intelligent output, like in 1974 translation would represent, does not mean you are reasoning in a human way.


> To me, Searle's point is ultimately that we don't know enough about the human mind to be able to judge whether it can be replicated artificially.

There's string circumstantial evidence that we do. And really "computers can simulate the physics in the brain" is the null hypothesis.

In any case why is the Chinese Room always stated as if it has a clear conclusion rather than "this doesn't really prove anything" if we don't know enough to say either way?

> And now that we have almost literally developed a Chinese Room

I don't think so. The GPTs are currently still very far from the complexity of the human brain, and they are missing many features that may make a big difference to consciousness - for instance the ability to learn while running.

So while it may be fairly easy to say ChatGPT isn't conscious/sentient, that isn't the question. It's whether computers theoretically can't be conscious because consciousness comes from some physical property that they can't reproduce (like quantum microtubule crankery).


> In any case why is the Chinese Room always stated as if it has a clear conclusion rather than "this doesn't really prove anything" if we don't know enough to say either way?

To me, it does have a clear and definitive conclusion, which is that mere intelligent output does not mean you are replicating human intelligence, or any higher order mechanism such as consciousness. We don't know enough to tell that it doesn't have any consciousness, but that's beside the point.

You mentioned that a computer that could simulate every molecule of a human brain would also likely replicate sentience. Of course the tricky part is how do you prove that assertion, if all you have is output? If I transfer your brain to an advanced computer as you describe, can I conclude that you're conscious based on what you tell me? I don't think so, because with present technology I could likely make a passable version of your writing output with a LLM. To me that's the real value of the Chinese Room, which is to expose precisely this dillemma. People wrote all sorts of replies to it in order to tackle that - you may be interested in reading about Dennett's p-zombies if you haven't already.


For me it's not an argument for or against sentience but a lot more about the difficulty of defining it and potentially measuring it in non-human systems along with the fuzziness of the concept of knowledge. My use of it was about my belief I guess that it's just pushing symbols around algorithmically without anything we'd call understanding.


I don't think it depends on that assumption. It merely asserts that replicating behavior that indicates sentience does not equate sentience, and today we can see that this is correct.


In general I’d agree but in regards to ChatGPT there is nothing about the way it works that suggests it is conscious or has a mind. If you asked an AI researcher to implement something akin to the way the program mentioned in the Chinese room argument works you’d pretty much get ChatGPT style solution.


Can you define sentience in this context?


It doesn't have a precise definition (which is part of the reason why the people saying programs fundamentally cannot be sentient are so stupid).

But basically it means being conscious / self-aware. Technically it means having some kind of senses that make you aware of the external environment too but that's a minor difference from consciousness - I only said "sentience" because it's what most other people talking about this say. They really mean consciousness. (And also text based IO can be a sense.)


It's not sentient, but it's not simply dropping the next likely token.

That doesn't explain why it can have a debate with itself as 4 distinct personalities at the same time, or act as a used car salesman and actually haggle with me over the price of a Ferrari, or write a story about anything you want, with a beginning, middle, and ending.

Those aren't in the corpus. After a tense negotion, I was able to get that Ferrari for $87,000 (ChatGPT originally wanted $120,000).

Emergent behavior is possible with these incredibly complex systems.


Not sure if you saw ChatGPT debate itself but it was pretty spectacularly crap? Of course, we know we know, ChatGPT 4 is absolutely mind mind blowing and better, but let's see it again soon.


No, I set it up to have a "quiz party" with itself, where ChatGPT was essentially the moderator asking the questions, and the 4 other "AIs" had distinct personalities, or heavy accents. Like one was angry and always thought it was right, another was nervous and unsure about it's answers, one was 5 years old and another could barely speak English with a heavy accent but was confident in its answers.

It was very interesting to watch. The "participants" actually started arguing with each other about the right answer, or straight-up insulting each other.


It's the multi-head attention that you're seeing in action. ChatGPT has 12, GPT-3 has 96.


Believing in "Chinese Room" non-sentience implies that you believe that there's something magical about flesh.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: