876

This is the technology worth trillions of dollars huh

top 50 comments
sorted by: hot top controversial new old
[-] mrductape@eviltoast.org 109 points 1 month ago

Well, it's almost correct. It's just one letter off. Maybe if we invest millions more it will be right next time.

Or maybe it is just not accurate and never will be....I will not every fully trust AI. I'm sure there are use cases for it, I just don't have any.

[-] TheFogan@programming.dev 28 points 1 month ago* (last edited 1 month ago)

Cases where you want something googled quickly to get an answer, and it's low consequence when the answer is wrong.

IE, say a bar arguement over whether that guy was in that movie. Or you need a customer service agent, but don't actually care about your customers and don't want to pay someone, or your coding a feature for windows.

[-] elvith@feddit.org 16 points 1 month ago

How it started:

Or you need a customer service agent, but don't actually care about your customers and don't want to pay someone

How it's going:

IKEA

Chevy

...

[-] mrductape@eviltoast.org 15 points 1 month ago

Chatbots are crap. I had to talk to one with my ISP when I had issues. Within one minute I had to request it to connect me to a real person. The problem I was having was not a standard issue, so of course the bot did not understand at all... And I don't need a bot to give me all the standard solutions, I've already tried all of that before I even contact customer support.

load more comments (1 replies)
load more comments (1 replies)
load more comments (3 replies)
[-] Djehngo@lemmy.world 60 points 1 month ago

The letters that make up words is a common blind spot for AIs, since they are trained on strings of tokens (roughly words) they don't have a good concept of which letters are inside those words or what order they are in.

[-] PixelatedSaturn@lemmy.world 12 points 1 month ago

I find it bizarre that people find these obvious cases to prove the tech is worthless. Like saying cars are worthless because they can't go under water.

[-] skisnow@lemmy.ca 88 points 1 month ago* (last edited 1 month ago)

Not bizarre at all.

The point isn't "they can't do word games therefore they're useless", it's "if this thing is so easily tripped up on the most trivial shit that a 6-year-old can figure out, don't be going round claiming it has PhD level expertise", or even "don't be feeding its unreliable bullshit to me at the top of every search result".

[-] PixelatedSaturn@lemmy.world 10 points 1 month ago

I don't want to defend ai again, but it's a technology, it can do some things and can't do others. By now this should be obvious to everyone. Except to the people that believe everything commercials tell them.

[-] kouichi@ani.social 23 points 1 month ago

How many people do you think know that AIs are "trained on tokens", and understand what that means? It's clearly not obvious to those who don't, which are roughly everyone.

load more comments (3 replies)
[-] sqgl@sh.itjust.works 18 points 1 month ago* (last edited 1 month ago)

358 instances (so far) of lawyers in Australia using AI evidence which "hallucinated".

And this week one was finally punished.

load more comments (4 replies)
load more comments (21 replies)
[-] knatschus@discuss.tchncs.de 20 points 1 month ago

Then why is Google using it for question like that?

Surely it should be advanced enough to realise it's weakness with this kind of questions and just don't give an answer.

[-] PixelatedSaturn@lemmy.world 14 points 1 month ago* (last edited 1 month ago)

They are using it for every question. It's pointless. The only reason they are doing it is to blow up their numbers.

... they are trying to be infront. So that some future ai search wouldn't capture their market share. It's a safety thing even if it's not working for all types of questions.

[-] echodot@feddit.uk 12 points 1 month ago

Well it also can't code very well either

load more comments (4 replies)
load more comments (8 replies)
load more comments (1 replies)
[-] ilinamorato@lemmy.world 59 points 1 month ago

✅ Colorado

✅ Connedicut

✅ Delaware

❌ District of Columbia (on a technicality)

✅ Florida

But not

❌ I'aho

❌ Iniana

❌ Marylan

❌ Nevaa

❌ North Akota

❌ Rhoe Islan

❌ South Akota

[-] individual@toast.ooo 30 points 1 month ago
load more comments (1 replies)
[-] boonhet@sopuli.xyz 12 points 1 month ago

Everyone knows it's properly spelled "I, the ho" not Idaho. That’s why it didn’t make the list.

[-] echodot@feddit.uk 52 points 1 month ago

You joke, but I bet you didn't know that Connecticut contained a "d"

I wonder what other words contain letters we don't know about.

[-] Rcklsabndn@sh.itjust.works 27 points 1 month ago

The famous 'invisible D' of Connecticut, my favorite SCP.

[-] kuberoot@discuss.tchncs.de 18 points 1 month ago

That actually sounds like a fun SCP - a word that doesn't seem to contain a letter, but when testing for the presence of that letter using an algorithm that exclusively checks for that presence, it reports the letter is indeed present. Any attempt to check where in the word the letter is, or to get a list of all letters in that word, spuriously fail. Containment could be fun, probably involving amnestics and widespread societal influence, I also wonder if they could create an algorithm for checking letter presence that can be performed by hand without leaking any other information to the person performing it, reproducing the anomaly without computers.

load more comments (5 replies)
load more comments (13 replies)
[-] Multiplexer@discuss.tchncs.de 38 points 1 month ago* (last edited 1 month ago)

No, this is Google throwing the cheapest possible shit at you that is barely capable of stringing together 5 coherent sentences and has the reasoning capability of a tapeworm.

Here is the output of the minimalist open Chinese model Qwen3, that runs locally on my 6 year old mid-end PC:

The US states that contain the letter "d" (case-insensitive, as state names are typically capitalized) are as follows. I've verified each state name for the presence of the letter "d" in its standard spelling:

    Colorado (contains "d" in "Colorado")  
    Delaware (starts with "D")  
    Florida (contains "d" in "Florida")  
    Idaho (contains "d" in "Idaho")  
    Indiana (contains "d" in "Indiana")  
    Maryland (ends with "d" in "Maryland")  
    Nevada (contains "d" in "Nevada")  
    North Dakota (contains "d" in "Dakota")  
    Rhode Island (contains "d" in "Rhode")  
    South Dakota (contains "d" in "Dakota")  


Total: 10 states.
load more comments (5 replies)
[-] dude@lemmings.world 31 points 1 month ago

Well, for anyone who knows a bit about how LLMs work, it’s pretty obvious why LLMs struggle with identifying the letters in the words

load more comments (13 replies)
[-] Blackmist@feddit.uk 30 points 1 month ago

Just another trillion, bro.

[-] NateNate60@lemmy.world 10 points 1 month ago

Just another 1.21 jigawatts of electricity, bro. If we get this new coal plant up and running, it'll be enough.

load more comments (1 replies)
[-] Jankatarch@lemmy.world 30 points 1 month ago* (last edited 1 month ago)

They took money away from cancer research programs to fund this.

[-] Burninator05@lemmy.world 13 points 1 month ago

After we pump another hundred trillion dollars and half the electricity generated globally into AI you're going to feel pretty foolish for this comment.

load more comments (1 replies)
load more comments (3 replies)
[-] sqgl@sh.itjust.works 30 points 1 month ago

ChatGPT is just as stupid.

[-] SaveTheTuaHawk@lemmy.ca 10 points 1 month ago

it's actually getting dumber.

[-] panda_abyss@lemmy.ca 28 points 1 month ago* (last edited 1 month ago)

Yesterday i asked Claude Sonnet what was on my calendar (since they just sent a pop up announcing that feature)

It listed my work meetings on Sunday, so I tried to correct it…

You’re absolutely right - I made an error! September 15th is a Sunday, not a weekend day as I implied. Let me correct that: This Week’s Remaining Schedule: Sunday, September 15

Just today when I asked what’s on my calendar it gave me today and my meetings on the next two thursdays. Not the meetings in between, just thursdays.

Something is off in AI land.

Edit: I asked again: gave me meetings for Thursday’s again. Plus it might think I’m driving in F1

[-] FlashMobOfOne@lemmy.world 18 points 1 month ago

A few weeks ago my Pixel wished me a Happy Birthday when I woke up, and it definitely was not my birthday. Google is definitely letting a shitty LLM write code for it now, but the important thing is they're bypassing human validation.

Stupid. Just stupid.

load more comments (1 replies)
load more comments (3 replies)
[-] kiku@feddit.org 26 points 1 month ago
[-] threeonefour@piefed.ca 28 points 1 month ago

Wait a sec, Minnasoda doesn't have a d??

load more comments (2 replies)
load more comments (2 replies)
[-] ArsonButCute@lemmy.dbzer0.com 25 points 1 month ago* (last edited 1 month ago)

Hey look the markov chain showed its biggest weakness (the markov chain)!

In the training data, it could be assumed by output that Connecticut usually follows Colorado in lists of two or more states containing Colorado. There is no other reason for this to occur as far as I know.

Markov Chain based LLMs (I think thats all of them?) are dice-roll systems constrained to probability maps.

Edit: just to add because I don't want anyone crawling up my butt about the oversimplification. Yes. I know. That's not how they work. But when simplified to words so simple a child could understand them, its pretty close.

[-] AlecSadler 12 points 1 month ago

Oh l I was thinking it's because people pronounce it Connedicut

load more comments (1 replies)
load more comments (1 replies)
[-] skisnow@lemmy.ca 22 points 1 month ago

I don't think this gets nearly enough visibility: https://www.academ-ai.info/

Papers in peer-reviewed journals with (extremely strong) evidence of AI shenanigans.

load more comments (2 replies)
[-] gilokee@lemmy.world 22 points 1 month ago
load more comments (2 replies)
[-] Arghblarg@lemmy.ca 20 points 1 month ago

"AI" hallucinations are not a problem that can be fixed in LLMs. They are an inherent aspect of the process and an inevitable result of the fact that LLMs are mostly probabilistic engines, with no supervisory or introspective capability, which actual sentient beings possess and use to fact-check their output. So there. :p

load more comments (4 replies)
[-] BlueMagma@sh.itjust.works 16 points 1 month ago

I get the sentiment behind this post, and it's almost always funny when LLM are such dumbass. But this is not a good argument against the technology. It is akin to climate change denier using the argument: "look! It snowed today, climate change is so dumb huh ?"

load more comments (10 replies)
[-] Aceticon@lemmy.dbzer0.com 15 points 1 month ago* (last edited 1 month ago)

"This is the technology worth trillions of dollars"

You can make anything fly high in the sky with enough helium, just not for long.

(Welcome to the present day Tech Stock Market)

load more comments (1 replies)
[-] SaveTheTuaHawk@lemmy.ca 13 points 1 month ago

We're turfing out students by the tens on academic misconduct. They are handing in papers with references that clearly state "generated by Chat GPT". Lazy idiots.

[-] NateNate60@lemmy.world 14 points 1 month ago

This is why invisible watermarking of AI-generated content is likely to be so effective. Even primitive watermarks like file metadata. It's not hard for anyone with technical knowledge to remove, but the thing with AI-generated content is that anyone who dishonestly uses it when they are not supposed to is probably also too lazy to go through the motions of removing the watermarking.

load more comments (7 replies)
load more comments (1 replies)
[-] samus12345@sh.itjust.works 13 points 1 month ago* (last edited 1 month ago)

Connedicut.

I wondered if this has been fixed. Not only has it not, the AI has added Nebraska.

load more comments (6 replies)
[-] Jaysyn@lemmy.world 11 points 1 month ago

Blows my mind people pay money for wrong answers.

[-] chaosCruiser@futurology.today 11 points 1 month ago* (last edited 1 month ago)

In Copilot terminology, this is a “quick response” instead of the “think deeper” option. The latter actually stops to verify the initial answer before spitting it out.

Deep thinking gave me this: Colorado, Delaware, Florida, Idaho, Indiana, Maryland, North Dakota, Rhode Island, and South Dakota.

It took way longer, but at least the list looks better now. Somehow it missed Nevada, so it clearly didn’t think deep enough.

[-] skisnow@lemmy.ca 25 points 1 month ago* (last edited 1 month ago)

"I asked it to burn an extra 2KWh of energy breaking the task up into small parts to think about it in more detail, and it still got the answer wrong"

load more comments (1 replies)
[-] Deestan@lemmy.world 10 points 1 month ago

Hey hey hey hey don't look at what it actually does.

Look at what it feels like it almost can do and pretend it soon will!

load more comments
view more: next ›
this post was submitted on 11 Sep 2025
876 points (100.0% liked)

Technology

76310 readers
2140 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS