460
you are viewing a single comment's thread
view the rest of the comments
[-] V0ldek@awful.systems 30 points 3 weeks ago

In my workflow there is no difference between LLMs and fucking grep for me.

Well grep doesn't hallucinate things that are not actually in the logs I'm grepping so I think I'll stick to grep.

(Or ripgrep rather)

[-] thedeadwalking4242@lemmy.world 17 points 3 weeks ago

With grep it’s me who hallucinates that I can right good regex :,)

[-] froztbyte@awful.systems 11 points 3 weeks ago

(I don't mean to take aim at you with this despite how irked it'll sound)

I really fucking hate how many computer types go "ugh I can't" at regex. the full spectrum of it, sure, gets hairy. but so many people could be well served by decently learning grouping/backrefs/greedy match/char-classes (which is a lot of what most people seem to reach for[0])

that said, pomsky is an interesting thing that might in fact help a lot of people go from "I want $x" as a human expression of intent, to "I have $y" as a regex expression

[0] - yeah okay sometimes you also actually need a parser. that's a whole other conversation. I'm talking about "quickly hacking shit up in a text editor buffer in 30s" type cases here

[-] swlabr@awful.systems 8 points 3 weeks ago* (last edited 3 weeks ago)

Hey. I can do regex. It's specifically grep I have beef with. I never know off the top of my head how to invoke it. Is it -e? -r? -i? man grep? More like, man, get grep the hell outta here!

[-] froztbyte@awful.systems 11 points 3 weeks ago

now listen, you might think gnu tools are offensively inconsistent, and to that I can only say

find(1)

[-] swlabr@awful.systems 12 points 3 weeks ago

find(1)? You better find(1) some other place to be, buster. In this house, we use the file explorer search bar

[-] lagoon8622@sh.itjust.works 5 points 3 weeks ago
[-] swlabr@awful.systems 4 points 3 weeks ago

If I start using this and add grep functionality to my day-to-day life, I can’t complain about not knowing how to invoke grep in good conscience, dawg. I can’t hold my shitposting back like that, dawg!

jk that looks useful. Thanks!

[-] lagoon8622@sh.itjust.works 4 points 3 weeks ago

The cheatsheet and tealdeer projects are awesome. It's one of my (many) favorite things about the user experience honestly. Really grateful for those projects

[-] thedeadwalking4242@lemmy.world 6 points 3 weeks ago

The funny thing is, I’m just going with the joke, I’m actually pretty good with regex lol

[-] froztbyte@awful.systems 4 points 3 weeks ago

woo! but still also check out pomsky, it's legit handy!

[-] froztbyte@awful.systems 3 points 3 weeks ago

(also I did my disclaimer at the start there, so, y'know (but also igwym))

[-] vivendi@programming.dev 1 points 3 weeks ago

Hallucinations become almost a non issue when working with newer models, custom inference, multishot prompting and RAG

But the models themselves fundamentally can't write good, new code, even if they're perfectly factual

[-] Architeuthis@awful.systems 17 points 3 weeks ago

If LLM hallucinations ever become a non-issue I doubt I'll be needing to read a deeply nested buzzword laden lemmy post to first hear about it.

load more comments (3 replies)
[-] scruiser@awful.systems 13 points 3 weeks ago

The promptfarmers can push the hallucination rates incrementally lower by spending 10x compute on training (and training on 10x the data and spending 10x on runtime cost) but they're already consuming a plurality of all VC funding so they can't 10x many more times without going bust entirely. And they aren't going to get them down to 0%, hallucinations are intrinsic to how LLMs operate, no patch with run-time inference or multiple tries or RAG will eliminate that.

And as for newer models... o3 actually had a higher hallucination rate because trying to squeeze rational logic out of the models with fine-tuning just breaks them in a different direction.

I will acknowledge in domains with analytically verifiable answers you can check the LLMs that way, but in that case, its no longer primarily an LLM, you've got an entire expert system or proof assistant or whatever that can operate independently of the LLM and the LLM is just providing creative input.

[-] swlabr@awful.systems 12 points 3 weeks ago

We should maximise hallucinations, actually. That is, we should hack the environmental controls of the data centers to be conducive for fungi growth, and flood them with magic mushrooms spores. We can probably get the rats on board by selling it as a different version of nuking the data centers.

[-] scruiser@awful.systems 12 points 3 weeks ago

What if [tokes joint] hallucinations are actually, like, proof the models are almost at human level man!

[-] swlabr@awful.systems 10 points 3 weeks ago

stopping this bit here because I don't want to continue writing a JRE episode

[-] pikesley@mastodon.me.uk 7 points 3 weeks ago
[-] froztbyte@awful.systems 8 points 3 weeks ago
[-] swlabr@awful.systems 8 points 3 weeks ago

Doesn't really narrow it down, sorry

(jk I don't have beef with the JavaRE)

[-] scruiser@awful.systems 7 points 3 weeks ago* (last edited 3 weeks ago)

Joe Rogan Experience!

...side note my most prominent irl conversation about Joe Rogan was with a relative who was trying to convince me it was a good thing that Joe Rogan platformed a celebrity who was saying 1x1=2 (Terrence Howard). Literally beyond parody.

[-] swlabr@awful.systems 6 points 3 weeks ago

Ok, we are already very far off topic. I actually have heard an interesting take about Terrence Howard and his “New Math”.

(NB: I have an undergrad major in mathematics)

So, one of my favourite comedy podcasts is “My Momma Told Me,” a podcast that talks about black conspiracy theories. In each episode, the topic is framed as “my momma told me <insert conspiracy theory here”. Terrence Howard comes up as a sort of mythological icon on the pod, it’s very funny. On a recent episode they actually get to facetime him, it rules.

Anyway, the take comes from the host Langston Kerman (who does not have a major in mathematics, nor any background in science). It’s a very charitable interpretation that the whole 1x1=2 stuff isn’t so much about creating a new math with different rules, it’s more an expression of the frustration towards the algorithms and calculations that are part of the power structures in society that disadvantage and disenfranchise minority populations. Because this is all being thought of and formulated by people outside of the discipline, anything appearing arithmetic is “math,” so a “new math” is needed.

Basically: it’s kafkaesque. But yeah more likely than not Terrence Howard has gone off the deep end.

[-] froztbyte@awful.systems 2 points 3 weeks ago

I hadn’t encountered either the Howard person nor heard of this podcast, but imma find that episode and listen because it sounds like quite an experience!

[-] MBM@lemmings.world 8 points 3 weeks ago

Sadly I have seen people make that exact point

load more comments (24 replies)
[-] kgMadee2@mathstodon.xyz 10 points 3 weeks ago

@vivendi @V0ldek * hallucinations are a fundamental trait of LLM tech, they're not going anywhere

[-] swlabr@awful.systems 9 points 3 weeks ago

God, this cannot be overstated. An LLM’s sole function is to hallucinate. Anything stated beyond that is overselling.

this post was submitted on 13 May 2025
460 points (100.0% liked)

TechTakes

1916 readers
78 users here now

Big brain tech dude got yet another clueless take over at HackerNews etc? Here's the place to vent. Orange site, VC foolishness, all welcome.

This is not debate club. Unless it’s amusing debate.

For actually-good tech, you want our NotAwfulTech community

founded 2 years ago
MODERATORS