220

Did nobody really question the usability of language models in designing war strategies?

you are viewing a single comment's thread
view the rest of the comments
[-] SlopppyEngineer@lemmy.world 31 points 1 year ago

LLM are just plagiarizing bullshitting machines. It's how they are built. Plagiarism if they have the specific training data, modify the answer if they must, make it up from whole cloth as their base programming. And accidentally good enough to convince many people.

[-] huginn@feddit.it 5 points 1 year ago

To be fair they're not accidentally good enough: they're intentionally good enough.

That's where all the salary money went: to find people who could make them intentionally.

[-] SlopppyEngineer@lemmy.world 7 points 1 year ago

GPT 2 was just a bullshit generator. It was like a politician trying to explain something they know nothing about.

GPT 3.0 was just a bigger version of version 2. It was the same architecture but with more nodes and data as far as I followed the research. But that one could suddenly do a lot more than the previous version, so by accident. And then the AI scene exploded.

[-] Limitless_screaming@kbin.social 2 points 1 year ago

It was the same architecture but with more nodes and data

So the architecture just needed more data to generate useful answers. I don't think that was an accident.

[-] Even_Adder@lemmy.dbzer0.com 4 points 1 year ago* (last edited 1 year ago)
[-] MNByChoice@midwest.social 3 points 1 year ago

I will read those, but I bet "accidentally good enough to convince many people." still applies.

A lot of things from LLM look good to nonexperts, but are full of crap.

[-] MNByChoice@midwest.social 2 points 1 year ago

https://notes.aimodels.fyi/self-rag-improving-the-factual-accuracy-of-large-language-models-through-self-reflection/

A cool paper. Using the LLM to judge value of new inputs.
I am always skeptical of summaries of journal articles. Even well meaning people can accidentally distort the conclusions.

Still LLM is a bullshit generator that can check bullshit level of inputs.

[-] MNByChoice@midwest.social 1 points 1 year ago

https://arxiv.org/abs/2310.02207

2 author paper with interesting evidence. Again, evidence not proof. Wait for the papers that cite this one.

[-] MNByChoice@midwest.social 1 points 1 year ago

https://notes.aimodels.fyi/researchers-discover-emergent-linear-strucutres-llm-truth/

References a 2 author paper. I am not an expert in the field, but it is important to read the papers that reference this one. Those papers will have criticisms that are thought out. In general, fewer authors means less debate between the authors and easier to miss details.

[-] MNByChoice@midwest.social 1 points 1 year ago

https://poke-llm-on.github.io/

Reinforcement learning. Cool project. Still no need to "know" anything. I usually play this type of have with short rules and monitoring the current state.

[-] MNByChoice@midwest.social 1 points 1 year ago* (last edited 1 year ago)

https://adamkarvonen.github.io/machine_learning/2024/01/03/chess-world-models.html

However, this only worked for a model trained on a synthetic dataset of games uniformly sampled from the Othello game tree. They tried the same techniques on a model trained using games played by humans and had poor results. To me, this seemed like a major caveat to the findings of the paper which may limit its real world applicability. We cannot, for example, generate code by uniformly sampling from a code tree.

Author later discusses training on you data versus general datasets.

I am out of my depth, but does not seem to provide strong evidence for the modem not just repeating information that shows up a lot for the given inputs.

[-] SlopppyEngineer@lemmy.world 1 points 1 year ago

Yes. There is self organization and possibility to self reflection going on in something that wasn't designed for it. That's going to spawn a lot more research.

[-] TrickDacy@lemmy.world 1 points 1 year ago

It kind of irks me how many people want to downplay this technology in this exact manner. Yes you're sort of right but in no way does that really change how it will be used and abused.

"But people think it's real AI tho!"

Okay and? Most people don't understand how most tech works and that doesn't stop it from doing a lot of good and bad things.

[-] SlopppyEngineer@lemmy.world 1 points 1 year ago

I've been through a few AI winters and hype cycles. It made me very cynical and convinced many overly enthusiastic people will run into a firewall face first.

this post was submitted on 05 Feb 2024
220 points (100.0% liked)

Technology

73331 readers
3217 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS