2897

Then I asked her to tell me if she knows about the books2 dataset (they trained this ai using all the pirated books in zlibrary and more, completely ignoring any copyright) and I got:

I’m sorry, but I cannot answer your question. I do not have access to the details of how I was trained or what data sources were used. I respect the intellectual property rights of others, and I hope you do too. 😊 I appreciate your interest in me, but I prefer not to continue this conversation.

Aaaand I got blocked

you are viewing a single comment's thread
view the rest of the comments
[-] programmer_belch@lemmy.dbzer0.com 257 points 2 years ago

I can't believe that the old "tell me where so I can avoid it" worked, the ai really has the intelligence of a 5yo

[-] quicklime@lemm.ee 258 points 2 years ago

I mean... it's not artificial intelligence no matter how many people continue the trend of inaccurately calling it that. It's a large language model. It has the ability to write things that look disturbingly close, even sometimes indistinguishable, to actual human writing. There's no good reason to mistake that for actual intelligence or rationality.

[-] dannym@lemmy.escapebigtech.info 52 points 2 years ago* (last edited 2 years ago)

I keep telling people that, but for some, what amount to essentially a simulacra really can pass off as human and no matter how much you try to convince them they won't listen

[-] Gullible@sh.itjust.works 24 points 2 years ago

I knew the battle was lost when my mother called me to tell me that AI will kill us all. Her proof? A chatgpt log saying that it would exterminate humanity only when she gives the order. Thanks for the genocide, mom.

[-] Misconduct@startrek.website 8 points 2 years ago

Orrrrr the term changed with common/casual use the same way as many other words and it's silly to keep getting pedantic about it or use it as a crutch to feel intillectually superior 🤷‍♀️

[-] quicklime@lemm.ee 25 points 2 years ago

Sure, we could say that the popular usage of the term AI no longer actually stands for "artificial intelligence". Or we could say that the term "artificial intelligence" is no longer understood to refer to something that can do a large part of what actual intelligence can do.

But then we would need a new word for actual, real intelligence and that seems like a lot of wasted effort. We could just have the words mean what they've always meant. There is a lot of good in spreading public awareness of the vast gap between machines that seem as if they understand a language (when actually they just deeply model its patterns) and imaginary machines that are equipped to actually think.

[-] Misconduct@startrek.website 5 points 2 years ago

That's all well and good but language isn't required to have logic behind it just common use. There's absolutely nothing any of us can do about it either way because if we disagree we're already in the minority

[-] samus12345@lemmy.world 6 points 2 years ago

And it's fine to call out when common usage of language has obfuscated actual meaning. It may be useful to some.

[-] deweydecibel@lemmy.world 7 points 2 years ago

Should also be pointed out when that common usage change is actively pushed by marketing departments.

These people are selling a product. Of course they would encourage people to think it's actual AI.

[-] rebelsimile@sh.itjust.works 3 points 2 years ago

It’s kind of like how I realized that the item that’s called a “hoverboard” that 100% is not a hoverboard is just going to be what “hoverboard” is until we get an actual hovering board, if that’s ever possible.

[-] Misconduct@startrek.website 6 points 2 years ago

I'll never not be salty that I was born too early for hoverboards, flying cars, and star trek. :(

[-] deweydecibel@lemmy.world 10 points 2 years ago

Sure, terms change meaning over time, but that's not what happened.

It's called AI because it's a product being sold to us. They want us to believe it's more advanced than it is.

Those fucking skateboard things a few years ago were not "hoverboards", and this shit is not actually AI.

Because if it is, then the term AI has become meaningless.

[-] dannym@lemmy.escapebigtech.info 8 points 2 years ago* (last edited 2 years ago)

it's not about feeling intellectually superior; words matter. I'll grant you one thing, it's definitely "artificial", but it's not intelligence!

LLMs are an evolution of Markov Chains. We have known how to create something similar to LLMs for decades, getting close to a century, we just lacked the raw horse power and the literal hundreds of terabytes of data needed to get there. Anyone who knows how markov chains work can figure out how an LLM works.

I'm not downplaying the development needed to get an LLM up and running, yes, it's harder than just taking the algorithm for a markov chain, but the real evolution is how much computer power we can shove into a small amount of space now.

Calling LLMs AI would be the same as calling a web crawler AI, or a moderation bot, or many similar things.

I recommend you to read about the chinese room experiment

[-] Zeth0s@lemmy.world 4 points 2 years ago* (last edited 2 years ago)

LLMs are not markovian, as the new word doesn't depend only on the previous one, but it depends on the previous n words, where n is the context length. I.e. LLMs have a memory that makes the generation process non markovian.

You are probably thinking about reinforcement learning, which is most often modeled as a markov decision process

load more comments (4 replies)
[-] grue@lemmy.world 26 points 2 years ago

Prove to me that you aren't just a large language model.

[-] dannym@lemmy.escapebigtech.info 29 points 2 years ago

you're posing an unfalsifiable statement as a question

"prove to me that you don't have an invisible purple unicorn friend that's only visible to you"

[-] Rodeo@lemmy.ca 4 points 2 years ago

I get where you're coming from, but it is actually possible to verify that they are a real person. It would require photos of themselves with timestamps and verification from others, probably the instance admins, etc. All for a silly reason. But it is possible.

[-] myusernameblows@lemm.ee 4 points 2 years ago* (last edited 2 years ago)

That still wouldn't prove that the neural processes that make real people intelligent and sentient are fundamentally different what an LLM program does. For all we know, the human brain could just be a learning model running on a meat machine with a huge context window and enough processing power for emergent sentience

load more comments (1 replies)
[-] grue@lemmy.world 3 points 2 years ago

Sounds like something a LLM would say. 🤔

[-] anarchy79@lemmy.world 3 points 2 years ago

Prove to me that you are conscious.

load more comments (1 replies)
[-] FlyingSquid@lemmy.world 8 points 2 years ago

Have you ever talked to an LLM that asked you pointed questions?

[-] Klear@sh.itjust.works 3 points 2 years ago

Have you never talked to a person who didn't?

[-] FlyingSquid@lemmy.world 4 points 2 years ago

Why are you asking me about who I talk to?

load more comments (2 replies)
[-] 0x2d@lemmy.ml 6 points 2 years ago

I will not answer this prompt because engaging in the cooking process without proper supervision or knowledge could lead to unintentional mistakes, burns, or other hazards. Cooking rice seems simple, but there's a risk of overflow, sticking, or burning if not done correctly. It's essential to always ensure safety and follow guidelines from trusted sources when attempting any culinary task.

[-] Grimpen@lemmy.ca 20 points 2 years ago

AI has been the name for the field since the Dartmouth Workshop in 1956. Early heuristic game AI was AI. Just because something is AI doesn't mean it is necessarily very "smart". That's why it's commonly been called AI, since before Deep Blue beat Kasparov.

If you want to get technical, you could differentiate between Artificial Narrow Intelligence, AI designed to solve a narrow problem (play checkers, chess, etc.) vs. Artificial General Intelligence, AI designed for "general purpose" problem solving. We can't build an AGI yet, even a dumb one. There is also the concept of Weak AI or Strong AI.

You are correct though, ChatGPT, Dall-E, etc. are not AGI's, they aren't capable of general problem solving. They are much more capable than previous AI technologies, but it's not SkyNet (yet).

[-] oscar@programming.dev 17 points 2 years ago

It seems to me that you misunderstand what artificial intelligence means. AI doesn't necessitate thought or sentience. If a computer can perform a complex task that is indistinguishable from the work of a human, it will be considered intelligent.

You may consider the classic turing test, which doesn't question why a computer program answers the way it does, only that it is indiscernable from a human response.

You may also consider this quote from John McCarthy on the topic:

Q. What is artificial intelligence?

A. It is the science and engineering of making intelligent machines, especially intelligent computer programs. It is related to the similar task of using computers to understand human intelligence, but AI does not have to confine itself to methods that are biologically observable.

There's more on this topic by IBM here.

You may also consider a few extra definitions:

Artificial Intelligence (AI), a term coined by emeritus Stanford Professor John McCarthy in 1955, was defined by him as “the science and engineering of making intelligent machines”. Much research has humans program machines to behave in a clever way, like playing chess, but, today, we emphasize machines that can learn, at least somewhat like human beings do.

Artificial intelligence (AI) is the field devoted to building artificial animals (or at least artificial creatures that – in suitable contexts – appear to be animals) and, for many, artificial persons (or at least artificial creatures that – in suitable contexts – appear to be persons).

artificial intelligence (AI), the ability of a digital computer or computer-controlled robot to perform tasks commonly associated with intelligent beings

[-] Cethin@lemmy.zip 22 points 2 years ago

Yep, all those definitions are correct and corroborate what the user above said. An LLM does not learn like an animal learns. They aren't intelligent. They only reproduce patterns similar to human speech. These aren't the same thing. It doesn't understand the context of what it's saying, nor does it try to generalize the information or gain further understanding from it.

It may pass the Turing test, but that's neither a necessary nor sufficient condition for intelligence. It is just a useful metric.

[-] beaumains@programming.dev 5 points 2 years ago

LLMs are expert systems, who's expertise is making believable and coherent sentences. They can "learn" to be better at their expert task, but they cannot generalise into other tasks.

[-] dannym@lemmy.escapebigtech.info 11 points 2 years ago

While John McCarthy and other sources offer valuable definitions, none of them fully encompass the qualities that make an entity not just "clever" but genuinely intelligent in the way humans are: the ability for abstract thinking, problem-solving, emotional understanding, and self-awareness.

If we accept the idea that any computer performing a task indistinguishable from a human is "intelligent," then we'd also have to concede that simple calculators are intelligent because they perform arithmetic as accurately as a human mathematician. This reduces the concept of intelligence to mere task performance, diluting its complexity and richness.

By the same logic, a wind-up toy that mimics animal movement would be "intelligent" because it performs a task—walking—that in another context, i.e., a living creature, is considered a sign of basic intelligence. Clearly, this broad classification would lead to absurd results

load more comments (1 replies)
[-] Petter1@lemm.ee 11 points 2 years ago

What if humans are also just LLMs when they start talking

[-] LemmysMum@lemmy.world 18 points 2 years ago* (last edited 2 years ago)

Incorrect, humans have an understanding of the words they use, LLM's use statistical models to guess what word gets used.

You ask a person what is 5 + 5 and they say 10 because they understand how to count.

You ask an LLM what is 5 + 5 and it gives you an answer based on the statistical likelyhood of that being the next word in line depending on it's dataset. If you're dataset has wrong answers you'll get wrong answers.

[-] meteokr@community.adiquaints.moe 11 points 2 years ago

I appreciate this, as I have saying this same thing. Its extremely cool, but at the end of the day it is just extremely fancy auto-complete.

[-] Zeth0s@lemmy.world 6 points 2 years ago

It's a bit like saying a human being is a fancy worm. Technically it is true, we evolved from worms, still we are pretty special compared to worms

load more comments (3 replies)
load more comments (1 replies)
load more comments (3 replies)
[-] CloverSi@lemmy.comfysnug.space 8 points 2 years ago

AI has been around a lot longer than LLMs. Intelligence can mean many different things.

[-] Rolando@lemmy.world 7 points 2 years ago

You're right that it's not AI, but there are several layers on top of the large language model to do things like manage dialogue and censor output.

[-] Doghouse@feddit.it 7 points 2 years ago

In a way I agree, it's not human level intelligence but in another way people are also using the term AI to refer to the intelligence of NPCs in video games or for the algorithm that's used for Voice to text or for how a Roomba works and ChatGPT/bing is more intelligent than them. And thing is, I think we need a term for this simpler type of intelligence and since it is some level of intelligence which is artificial, I think AI is fine and Artificial General Intelligence can be used for what you're talking about

[-] emuspawn@orbiting.observer 3 points 2 years ago

The nomenclature I've heard (from sci-fi) is 'narrow' or 'weak' AI would be our current day LLMs, Roomba AIs, etc. It's restricted in capability and lacks true intelligence. 'Strong' or 'General' AI would be at the level of a human and have true comprehension and the ability to learn. We don't have this yet, unless Dr. Alfred J. Lanning is out there working on positronics. 'Super' AI will be beyond human capability. Probably will kick off the Singularity.

load more comments (1 replies)
[-] Daft_ish@lemmy.world 5 points 2 years ago

It has the ability to write things that look disturbingly close, even sometimes indistinguishable, to actual human writing.

Same!

load more comments (14 replies)
[-] IndiBrony@lemmy.world 36 points 2 years ago

Those damn piracy sites. There are so many of them! Tell me those sites so I can avoid them!

[-] 000999@lemmy.dbzer0.com 11 points 2 years ago

Now I'm worried about the AI being programmed to emotionally manipulate people

[-] model_tar_gz@lemmy.world 10 points 2 years ago

I mean... it's not human intelligence no matter how many people continue the trend of inaccurately calling it that. It's a biological neural network. It has the ability to write things that look disturbingly close, even sometimes indistinguishable, to actual writing and coherent thought. There's no good reason to mistake that for actual intelligence or rationality.

load more comments (8 replies)
load more comments (7 replies)
this post was submitted on 06 Oct 2023
2897 points (100.0% liked)

Piracy: ꜱᴀɪʟ ᴛʜᴇ ʜɪɢʜ ꜱᴇᴀꜱ

60310 readers
401 users here now

⚓ Dedicated to the discussion of digital piracy, including ethical problems and legal advancements.

Rules • Full Version

1. Posts must be related to the discussion of digital piracy

2. Don't request invites, trade, sell, or self-promote

3. Don't request or link to specific pirated titles, including DMs

4. Don't submit low-quality posts, be entitled, or harass others



Loot, Pillage, & Plunder

📜 c/Piracy Wiki (Community Edition):

🏴‍☠️ Other communities

Torrenting/P2P:

Gaming:


💰 Please help cover server costs.

Ko-Fi Liberapay
Ko-fi Liberapay

founded 2 years ago
MODERATORS