258
top 50 comments
sorted by: hot top controversial new old
[-] swlabr@awful.systems 86 points 3 months ago

LLMs, and everyone who uses them to process information:

[-] hex@programming.dev 63 points 3 months ago

Facts are not a data type for LLMs

I kind of like this because it highlights the way LLMs operate kind of blind and drunk, they're just really good at predicting the next word.

[-] CleoTheWizard@lemmy.world 28 points 3 months ago

They’re not good at predicting the next word, they’re good at predicting the next common word while excluding most unique choices.

What results is essentially if you made a Venn diagram of human language and only ever used the center of it.

[-] hex@programming.dev 15 points 3 months ago

Yes, thanks for clarifying what I meant! AI will never create anything unique unless prompted uniquely and even then it will tend to revert back to what you expect most.

[-] swlabr@awful.systems 47 points 3 months ago

ATTN: If you're coming into this thread to say, "The output of AI is bad because your prompts suck," I'm just proud that you managed to figure out how to use the internet at all. Good job, you!

[-] froztbyte@awful.systems 15 points 3 months ago

remember remember, eternal september

(not that I much agree with the classist overtones of the original, but fuck me does it come to mind often)

[-] Sibbo@sopuli.xyz 34 points 3 months ago

Well, to be fair, AI can do it in seconds. Which beats humans.

But if that is relevant if the results are worthless is another question.

[-] HubertManne@moist.catsweat.com 13 points 3 months ago

Yeah it changes the task from note taking or summarizing to proofreading.

load more comments (3 replies)
[-] RagnarokOnline@programming.dev 27 points 3 months ago

I had GPT 3.5 break down 6x 45-minute verbatim interviews into bulleted summaries and it did great. I even asked it to anonymize people’s names and it did that too. I did re-read the summaries to make sure no duplicate info or hallucinations existed and it only needed a couple of corrections.

Beats manually summarizing that info myself.

Maybe their prompt sucks?

[-] froztbyte@awful.systems 41 points 3 months ago

“Are you sure you’re holding it correctly?”

christ, every damn time

load more comments (6 replies)
[-] dgerard@awful.systems 29 points 3 months ago

I got AcausalRobotGPT to summarise your post and it said "I'm not saying it's always programming.dev, but"

[-] pikesley@mastodon.me.uk 24 points 3 months ago

@RagnarokOnline @dgerard "They failed to say the magic spells correctly"

[-] HootinNHollerin@lemmy.world 18 points 3 months ago

Did you conduct or read all the interviews in full in order to verify no hallucinations?

load more comments (5 replies)
[-] kbal@fedia.io 22 points 3 months ago

Made strange choices about what to highlight.

They certainly do. For a while it was common to see AI-generated summaries under links to articles on lemmy, so I got a feel for them. Seems to me you would not need any fancy artificial intelligence to do equally well: Just take random excerpts, or maybe just read every third sentence.

[-] dgerard@awful.systems 21 points 3 months ago

how the hell did this of all the posts turn into a promptfondler shooting gallery

[-] froztbyte@awful.systems 11 points 3 months ago

1.26K subscribers

load more comments (2 replies)
[-] dgerard@awful.systems 19 points 3 months ago

i have seen the light from the helpful posters here, made up bullshit alleged summaries of documents are great actually

[-] khalid_salad@awful.systems 11 points 3 months ago

Could it be because a statistical relation isn't the same as a semantic one? No, I must be prompting it wrong. I'll just add "engineer" to my title and then everyone will take me seriously.

[-] lvxferre@mander.xyz 10 points 3 months ago* (last edited 3 months ago)

You could use them to know what the text is about, and if it's worth your reading time. In this situation, it's fine if the AI makes shit up, as you aren't reading its output for the information itself anyway; and the distinction between summary and shortened version becomes moot.

However, here's the catch. If the text is long enough to warrant the question "should I spend my time reading this?", it should contain an introduction for that very purpose. In other words if the text is well-written you don't need this sort of "Gemini/ChatGPT, tell me what this text is about" on first place.

EDIT: I'm not addressing documents in this. My bad, I know. [In my defence I'm reading shit in a screen the size of an ant.]

[-] queermunist@lemmy.ml 21 points 3 months ago* (last edited 3 months ago)

ChatGPT gives you a bad summary full of hallucinations and, as a result, you choose not to read the text based on that summary.

load more comments (13 replies)
[-] dgerard@awful.systems 20 points 3 months ago

Both the use cases here are goverment documents. I'm baffled at the idea of it being "fine if the AI makes shit up".

load more comments (1 replies)
load more comments (4 replies)
[-] beefbot 9 points 3 months ago

Is it only me, or is the linked article not super long on details & is reaching a conclusion from 2 examples? This is important & I need to hear more, & I’m generally biased against AI at this point— but the article isn’t doing enough to convince me

[-] self@awful.systems 14 points 3 months ago

did you click through to any of the inline citations? David’s shorter articles on pivot mostly gather and summarize those, so if you need to read the original research and its conclusions that’s where to go

[-] beefbot 11 points 3 months ago

Ah, that’s better, yes. Thank you , no sarcasm :) now sleepy brain is more informed

load more comments
view more: next ›
this post was submitted on 04 Sep 2024
258 points (100.0% liked)

TechTakes

1489 readers
50 users here now

Big brain tech dude got yet another clueless take over at HackerNews etc? Here's the place to vent. Orange site, VC foolishness, all welcome.

This is not debate club. Unless it’s amusing debate.

For actually-good tech, you want our NotAwfulTech community

founded 2 years ago
MODERATORS