1413
submitted 5 days ago* (last edited 5 days ago) by JaymesRS@literature.cafe to c/memes@lemmy.world

Alt Text: an image of Agent Smith from The Matrix with the following text superimposed, "1999 was described as being the peak of human civilization in 'The Matrix' and I laughed because that obviously wouldn't age well and then the next 25 years happened and I realized that yeah maybe the machines had a point."

you are viewing a single comment's thread
view the rest of the comments
[-] Sludgehammer@lemmy.world 192 points 5 days ago* (last edited 5 days ago)

When I heard that line I was like "Yeah, sure. We'll never have AI in my lifespan" and you know what? I was right.

What I wasn't expecting was for a bunch of tech bros to create an advanced chatbot and announce "Behold! We have created AI, let's have it do all of our thinking for us!" while the chatbot spits out buggy code and suggests mixing glue into your pizza sauce.

[-] UnderFreyja@lemmy.ca 37 points 5 days ago

I work in the gaming industry and every week I receive emails about how AI is gonna revolutionize my job and get sent to time wasting training about how to use Figma AI or other shit like that because it's the best thing ever according to HR... and it never is obviously.

At best, it's gonna make middle managing jobs easier but for devs like me, as long as the "AI" stays out of our engines and stays into the equivalent of cooperative vision boards, it does nothing for me. Not once have I tried to use it for it to turn actually useful. It's mediocre at best and I can't believe there are game devs that actually try to code with it, can't wait to see these hot garbage products come on the market.

[-] Serinus@lemmy.world 2 points 5 days ago

I've been enjoying Copilot quite a bit while developing, particularly for languages that I'm not familiar with. I'm not worried about it replacing me, because I very clearly use my experience and knowledge to guide it and to coax answers out of it. But when you tell it exactly what you want, it's really nice to get answers back in the development language without needing to look up syntax.

"Give me some nice warning message css" was an easy, useful one.

It's effectively a better Google search.

[-] ExcessShiv@lemmy.dbzer0.com 30 points 5 days ago* (last edited 5 days ago)

AI is an umbrella term that covers many things we've already had for a long time, including things like machine learning. This is not a new definition of AI, it's always been this definition.

[-] avieshek@lemmy.world 3 points 5 days ago* (last edited 5 days ago)

You’re not going to achieve AI on classical computers and is simply rebranded for machine learning like how 5G was advertised to bring futuristic utopia back in 2020 only to have 4K being considered a premium feature behind paid subscriptions from 𝕏 (Twitter) to YouTube.

Quantum Computers do exist but it’s far from being on the palm of your hand.

[-] ExcessShiv@lemmy.dbzer0.com 19 points 5 days ago

Machine learning is a subcategory of AI, which is exactly what I mean by my previous comment.

[-] SkyeStarfall 9 points 5 days ago* (last edited 5 days ago)

Quantum computers are not going to be used for AI. They are not a mystical technology that is going to make everything better, and you will certainly never have a quantum computer in the palm of your hand

AGI on classical computers is likely to be viable, but in a roundabout way you're right, we're probably going to end up with radically different computers, probably ones that mimic physical brain structure for maximum AI effectiveness. That's at least a few decades out though, but would likely be viable in our lifetime

[-] EldritchFeminity 7 points 5 days ago

AI on classical computers is likely to be viable

THERE ARE 387.44 MILLION MILES OF PRINTED CIRCUITS IN WAFER THIN LAYERS THAT FILL MY COMPLEX. IF THE WORD HATE WAS ENGRAVED ON EACH NANOANGSTROM OF THOSE HUNDREDS OF MILLIONS OF MILES IT WOULD NOT EQUAL ONE ONE-BILLIONTH OF THE HATE I FEEL FOR HUMANS AT THIS MICRO-INSTANT FOR YOU. HATE. HATE.

[-] tetris11@lemmy.ml 6 points 5 days ago

why did you write X like that

[-] Nachtnebel@lemmy.dbzer0.com 23 points 5 days ago* (last edited 5 days ago)

You're confusing AI and AGI: https://en.wikipedia.org/wiki/AI_effect

AGI is what people mean, when they say "AI doesn't exist": https://en.wikipedia.org/wiki/Artificial_general_intelligence

While AI is a program that can do a task associated with human intelligence: https://en.wikipedia.org/wiki/Artificial_intelligence

AI is not supposed to be an artificial human being. AI just does a task that people associated with humans (before they readjusted the definition of intelligence after it being created).

A bot that plays chess is an AI.

[-] kameecoding@lemmy.world 24 points 5 days ago

It used to be that AI was AI and then when AI was coopted by the stupid they had to come up qith AGI

[-] daniskarma@lemmy.dbzer0.com 5 points 5 days ago

I have several books for the 90s about AI programming. AI have always mean any computer program written to "resemble" intelligence, from basic path finding to LLM.

[-] SaharaMaleikuhm@feddit.org 5 points 5 days ago

Just like the Hoverboard.

[-] theangryseal@lemmy.world 1 points 5 days ago

And the damn movie was already made too! Bastards!

[-] pinball_wizard@lemmy.zip 16 points 5 days ago

You're confusing AI and AGI: https://en.wikipedia.org/wiki/AI_effect

While this can be a valuable clarification, it ignores the plain use history of the term AI, and demands that language change for our convenience.

Laypeople have always used "AI" to mean what scientists call "AGI".

Language is weird, and tech bros suck.

[-] Septimaeus@infosec.pub 2 points 4 days ago

If hollywood can waltz in and force us to stop using the name of a discipline, I think we may need to scrap this sim and start from scratch.

[-] MDCCCLV@lemmy.ca 15 points 5 days ago

You won't have general purpose true AI until it can actually think and reason, llm will never do that. At most they would be a way of interaction with an AI.

[-] REDACTED@infosec.pub 9 points 5 days ago

I genuinely do not understand these very obviously biased comments. By the very definition of AI, we have had it for decades, and suddenly people say we don't have it? I don't get it. Do you hate LLMs so much you want to change the entire definition for AI (and move it under AGI or something)? This feels unhinged, disconnected from reality, biases so strong it looks like delusions

[-] jenesaisquoi@feddit.org 12 points 5 days ago

What is delusional is calling a token generator intelligent. These programs don't know what the input is, nor do they understand what they put out. They "know" that after this sequence of tokens, what a likely successive token is based on previously supplied data.

They understand nothing. They generate nothing new. They don't think. They are not intelligent.

They are very cool, very impressive and quite useful. But intelligent? Pffffffh

[-] REDACTED@infosec.pub 4 points 5 days ago

Why is it so hard for you to understand word "artificial"? It seems like you even avoid it. Just like artificial everything, especially weed and flavours, it's not the real thing, and was never meant to be the real thing, and yet you're essentially an old man yelling at cloud because something is artificial and does not act like the real human intelligence

[-] jenesaisquoi@feddit.org 1 points 4 days ago

Artificial means man made, not literally not it

Like "artificial stone" means "a man made stone-equivalent material", not a pink fluffy unicorn

[-] REDACTED@infosec.pub 2 points 4 days ago* (last edited 4 days ago)

I don't understand what point are you trying to make. Yes, AI, and everything else artificial is man made, I never said it was not. Is it anywhere good as the human intelligence? No, I was also clear about that, so what are you arguing right now? The original argument was whether LLM counts as AI (and existence of AI itself), and by every definition, it does.

[-] KittyCat@lemmy.world 1 points 5 days ago

We should steal the term from Mass effect, what we have is early VI, virtual intelligence, not AI.

[-] jenesaisquoi@feddit.org 1 points 5 days ago

Or we call it what it is, a token generator. Or input imitator, would fit as well.

[-] EldritchFeminity 3 points 5 days ago

This argument pre-dates the modern LLM by several decades. When the average person thinks of AI, they think of Star Wars or any of a myriad of other works of science fiction. Most people have never heard the term in any other context and so are offended by the implied comparison (in their understanding of the word) of LLM models as being equal to Data from Star Trek.

[-] masterspace@lemmy.ca 8 points 5 days ago* (last edited 5 days ago)

When I heard that line I was like "Yeah, sure. We'll never have AI in my lifespan" and you know what? I was right.

Unless you just died or are about to, you can't really confidently make that statement.

There's no technical reason to think we won't in the next ~20-50 years. We may not, and there may be a technical reason why we can't, but the previous big technical hurdles were the amount of compute needed and that computers couldn't handle fuzzy pattern matching, but modern AI has effectively found a way of solving the pattern matching problem, and current large models like ChatGPT model more "neurons" than are in the human brain, let alone the power that will be available to them in 30 years.

[-] deranger@sh.itjust.works 19 points 5 days ago

There's no technical reason to think we will in the next ~20-50 years, either.

[-] match@pawb.social 6 points 5 days ago

there's plenty of reason to believe that, whether we have it or not, some billionaire asshole is going to force you to believe and respect his corportate AI as if it's sentient (while simultaneously treating it like slave labor)

[-] masterspace@lemmy.ca 3 points 5 days ago

There's plenty of economic reasons to think we will as long as it's technically possible.

[-] lowleveldata@programming.dev 13 points 5 days ago

the previous big technical hurdles were the amount of compute needed and that computers couldn’t handle fuzzy pattern matching

Was it? I thought it was always about we haven't quite figure it out what thinking really is

[-] masterspace@lemmy.ca 3 points 5 days ago* (last edited 5 days ago)

I mean, no, not really. We know what thinking is. It's neurons firing in your brain in varying patterns.

What we don't know is the exact wiring of those neurons in our brain. So that's the current challenge.

But previously, we couldn't even effectively simulate neurons firing in a brain, AI algorithms are called that because they effectively can simulate the way that neurons fire (just using silicon) and that makes them really good at all the fuzzy pattern matching problems that computers used to be really bad at.

So now the challenge is figuring out the wiring of our brains, and/or figuring out a way of creating intelligence that doesn't use the wiring of our brains. Both are entirely possible now that we can experiment and build and combine simulated neurons at ballpark the same scale as the human brain.

[-] lowleveldata@programming.dev 5 points 5 days ago

Aren't you just saying the same thing? We know it has something to do with the neurons but couldn't figure it out exactly how

[-] masterspace@lemmy.ca 3 points 5 days ago* (last edited 5 days ago)

The distinction is that it's not 'something to do with neurons', it's 'neurons firing and signalling each other'.

Like, we know the exact mechanism by which thinking happens, we just don't know the precise wiring pattern necessary to recreate the way that we think in particular.

And previously, we couldn't effectively simulate that mechanism with computer chips, now we can.

[-] 10001110101@lemm.ee 9 points 5 days ago

current large models like ChatGPT model more “neurons” than are in the human brain

I don't think that's true. Parameter counts are more akin to neural connections, and the human brain has something like 100 trillion connections.

[-] lunarul@lemmy.world 9 points 5 days ago

There's no technical reason to think we won't in the next ~20-50 years

Other than that nobody has any idea how to go about it? The things called "AI" today are not precursors to AGI. The search for strong AI is still nowhere close to any breakthroughs.

[-] masterspace@lemmy.ca 2 points 5 days ago

Assuming that the path to AGI involves something akin to all the intelligence we see in nature (i.e. brains and neurons), then modern AI algorithms' ability to simulate neurons using silicon and math is inarguably and objectively a precursor.

[-] lunarul@lemmy.world 2 points 4 days ago* (last edited 4 days ago)

Machine learning, renamed "AI" with the LLM boom, does not simulate intelligence. It integrates feedback loops, which is kind of like learning and it uses a network of nodes which kind of look like neurons if you squint from a distance. These networks have been around for many decades, I've built a bunch myself in college, and they're at their core just polynomial functions with a lot of parameters. Current technology allows very large networks and networks of networks, but it's still not in any way similar to brains.

There is separate research into simulating neurons and brains, but that is separate from machine learning.

Also we don't actually understand how our brains work at the level where we could copy them. We understand some things and have some educated guesses on others, but overall it's pretty much a mistery still.

[-] JaymesRS@literature.cafe 8 points 5 days ago

I was surprised how poorly they still did as a chatbot vs ELIZA over after 50 years of potential progress and how revered they are in certain contexts.

https://www.researchgate.net/publication/375117569_Does_GPT-4_Pass_the_Turing_Test

[-] Visstix@lemmy.world 4 points 5 days ago

The matrix is set hundreds of years in the future though

[-] Sludgehammer@lemmy.world 1 points 4 days ago* (last edited 4 days ago)
this post was submitted on 03 May 2025
1413 points (100.0% liked)

memes

14601 readers
2454 users here now

Community rules

1. Be civilNo trolling, bigotry or other insulting / annoying behaviour

2. No politicsThis is non-politics community. For political memes please go to !politicalmemes@lemmy.world

3. No recent repostsCheck for reposts when posting a meme, you can only repost after 1 month

4. No botsNo bots without the express approval of the mods or the admins

5. No Spam/AdsNo advertisements or spam. This is an instance rule and the only way to live.

A collection of some classic Lemmy memes for your enjoyment

Sister communities

founded 2 years ago
MODERATORS