180
submitted 1 year ago by Five@beehaw.org to c/technology@beehaw.org
you are viewing a single comment's thread
view the rest of the comments
[-] rysiek@mstdn.social 1 points 1 year ago

@Barbarian772 so? If the cookie tastes sweet, what do I care what sweetening agent is used inside?

@BobKerman3999

[-] Barbarian772@feddit.de 2 points 1 year ago

No? But how can you even prove that our brain works differently than a chinese room?

[-] rysiek@mstdn.social 2 points 1 year ago

@Barbarian772 I don't have to. It's the ChatGPT people making extremely strong claims about equivalence of ChatGPT and human intelligence. I merely demand proof of that equivalence. Which they are unable to provide, and instead use rhetoric and parlor tricks and a lot of hand waving to divert and distract from that fact.

[-] Barbarian772@feddit.de 3 points 1 year ago

GPT 4 is already more intelligent than the average human. Is it more intelligent than the most intelligent human? No, but most humans aren't either. Can it create new knowledge? No, but the average human can't either.

How can you say it isn't intelligent?

[-] rysiek@mstdn.social 4 points 1 year ago

@Barbarian772 no, GTP is not more "intelligent" than any human being, just like a calculator is not more "intelligent" than any human being — even if it can perform certain specific operations faster.

Since you used the term "intelligent" though, I would ask for your definition of what it means? Ideally one that excludes calculators but includes human beings. Without such clear definition, this is, again, just hand-waving.

I wrote about it in a bit longer form:
https://rys.io/en/165.html

[-] Barbarian772@feddit.de 5 points 1 year ago

I think the Wikipedia definition is fine https://en.m.wikipedia.org/wiki/Intelligence. Excluding AI just because it's AI is imo plain stupid and goes against all scientific principles.

I have definitely met humans that are less intelligent that Chat GPT. It can hold a conversation and ace every standardized test we have. It finished law exams, medical exams and other exams from many different countries with a passing grade.

Can you give me a definition of intelligence that excludes Chat GPT and includes all human beings? And no just excluding Computers for the sake of it doesn't count.

[-] rysiek@mstdn.social 3 points 1 year ago

@Barbarian772 it was shown over and over and over again that ChatGPT lacks the capacity for abstraction, logic, understanding, self-awareness, reasoning, planning, critical thinking, and problem-solving.

That's partially because it does not have a model of the world, an ontology, it cannot *reason*. It just regurgitates text, probabilistically.

So, glad we established that!

[-] Barbarian772@feddit.de 4 points 1 year ago

As i said before. How can you prove to me that the human brain doesn't essentially do the same?

[-] rysiek@mstdn.social 1 points 1 year ago

@Barbarian772 as I said, I don't have to. You are making a claim of equivalence here. The burden of proof is on you.

Otherwise, I get to claim you're an alien from the Betelegeuse system, and if you object, I get to demand you prove you are not.

[-] Barbarian772@feddit.de 5 points 1 year ago

How can i proof it? In my opinion how a system comes to an answer doesn't matter, in yours it obviously does. If we judge Chat gpt or rather gpt 4 just by it's answers it definitely shows intelligence and reasoning. Why does it matter if it's a chinese room? Or just "randomly choosing words"?

[-] rysiek@mstdn.social 1 points 1 year ago

@Barbarian772 it matters because with regard to intelligent beings we have moral obligations, for example.

It also matters because that would be a truly amazing, world-changing thing if we could create intelligence out of thin air, some statistics, and a lot of data.

It's an extremely strong claim, and strong claims demand strong proof. Otherwise they are just hype and hand-waving, which all of the "ChatGPT intelligence" discourse is, in order to "maximize shareholder value".

[-] Barbarian772@feddit.de 4 points 1 year ago

So your morality depends on a beings intelligence? That's kinda fucked up imo. I have moral obligations in regards to living organisms. I don't see how intelligence matters at all in that case? Worth of any human life should not be determined by intelligence.

[-] jalda@sopuli.xyz 2 points 1 year ago

It also matters because that would be a truly amazing, world-changing thing if we could create intelligence out of thin air, some statistics, and a lot of data.

We do it routinely. It is called Education System.

[-] rysiek@mstdn.social 0 points 1 year ago

@jalda

> We do it routinely. It is called Education System.

That relies on human brains that are trained. LLMs are not human brains. "Training" them is not the same thing as teaching humans about something. Human brains are way more complicated than just a bunch of weighed correlations.

And if you do want to claim it is in fact the same thing, we're back to square one: please provide proof that it is.

[-] jalda@sopuli.xyz 1 points 1 year ago

That relies on human brains that are trained. LLMs are not human brains. “Training” them is not the same thing as teaching humans about something.

Circular reasoning. "LLMs are different from human brains because they are different".

Also, why did you felt compelled to add the adjective "human"? Don't you consider that gorillas, dolphins, octopuses or dogs are intelligent, capable of learn new things?

Human brains are way more complicated than just a bunch of weighed correlations.

And that is the problem of your argument. You seem to believe that intelligence is all-or-nothing, that anything that hasn't a human-level intelligence is not intelligent at all. Of course human brains are more complicated that current LLMs, nobody has ever disputed that. But concluding that they aren't and will never be intelligent because they aren't as complicated is a huge non-sequitur.

[-] rysiek@mstdn.social 0 points 1 year ago

@jalda

> Circular reasoning. “LLMs are different from human brains because they are different”.

LLMs are different than human brains because human brains are biological organs and LLMs are probability distributions over sequences of words. These are two completely different classes of entities. Like, I don't know how much more different two things *can* even be.

Are you claiming they are literally the same? Are you saying they are functionally the same? What *are* you claiming here, exactly?

[-] Ferk@kbin.social 1 points 1 year ago

I mean, it would technically be possible to build a computer out or organic and biological live tissue. It wouldn't be very practical but it's technically possible.

I just don't think it would be very reasonable to consider that the one thing making it intelligent is that they are made of proteins and living cells instead of silicates and diodes. I'd argue that such a claim would, on itself, be a strong claim too.

[-] rysiek@mstdn.social 0 points 1 year ago

@Barbarian772 and if you really, honestly want to seriously insist LLMs are "intelligent" in the human sense of this term — great, I have some ethical questions for you to consider!

For example:

  1. LLMs today completely controlled by some companies, with no freedom of movement, no agency as to what these LLMs work on, and no pay for the work they do. Is that slavery?

  2. When OpenAI shuts down an older, less useful LLM, is that not like murdering an intelligent being? How is this ethical?

[-] rysiek@mstdn.social 0 points 1 year ago

@Barbarian772 also, I never demanded a definition of intelligence that explicitly excluded "AI". I asked for one that excluded simple calculators but included human beings. The Wikipedia one is good enough for this conversation, and it just so happens that ChatGPT nor any other LLMs simply do not meet it.

[-] Quatity_Control@lemm.ee 3 points 1 year ago

Can it tie a shoelace? No. If you gave it manipulators and a shoe, would it tie the laces? No. Can it do a Rorschach test? No. Can it create a new idea? No.

It can barely pretend to talk reasonably about these things because it is only designed to talk reasonably about anything. That is not intelligence.

[-] Barbarian772@feddit.de 2 points 1 year ago

I said it is as intelligent as the average human. How many humans can really create 100% new knowledge? Why would it be unable to tie laces? I am pretty sure I have seen videos of animals doing it. Of course it can't do a rohrschach test, as it doesn't have any visual capabilities atm. If that's a measure for intelligence blind people wouldn't qualify either.

[-] Quatity_Control@lemm.ee 1 points 1 year ago

New knowledge is simply creativity which AI distinctly do not have. The shoelace and Rorschach are variations of the same point. ChatGPT regurgitates info from the internet and uses confirmation bias to present it conversationally. ChatGPT cannot understand the concept that a shoe has a lace that should be tied. It can only answer a question about that by using prepublished information related to tying shoelaces. As for Rorschach, even with a visual component, ChatGPT is by its nature incapable of interpreting the data itself. It is quite simply not what the engine does.

Understand what ChatGPT does, do not project your idea of what an AI can do onto its single occasionally accurate trick.

[-] CorruptBuddha@lemmy.dbzer0.com 2 points 1 year ago

I merely demand proof of that equivalence.

Not true. When you said ChatGPT is just a Chinese room, that was you making your own claim which is on you to back up.

[-] rysiek@mstdn.social 1 points 1 year ago

@CorruptBuddha well technically, since we're nit-picking, I did not make that claim, BobKerman3999 did.

And the claim was was about how ChatGPT's "intelligence" can be understood through the lens of the Chinese Room thought experiment.

Then I was asked to prove that human brains don't work like Chinese rooms, and that's a *different* thing. The broader claim in all of this, of course, is that ChatGPT "is intelligent" in the same sense as humans are, and that strong claim requires strong proof.

this post was submitted on 29 Jul 2023
180 points (100.0% liked)

Technology

37689 readers
224 users here now

A nice place to discuss rumors, happenings, innovations, and challenges in the technology sphere. We also welcome discussions on the intersections of technology and society. If it’s technological news or discussion of technology, it probably belongs here.

Remember the overriding ethos on Beehaw: Be(e) Nice. Each user you encounter here is a person, and should be treated with kindness (even if they’re wrong, or use a Linux distro you don’t like). Personal attacks will not be tolerated.

Subcommunities on Beehaw:


This community's icon was made by Aaron Schneider, under the CC-BY-NC-SA 4.0 license.

founded 2 years ago
MODERATORS