264
submitted 1 year ago by misk@sopuli.xyz to c/technology@lemmy.world
you are viewing a single comment's thread
view the rest of the comments
[-] seiryth@lemmy.world 121 points 1 year ago

The thing that shits me about this is google appear to the public to be late to the party but the reality is they DID put safety before profit when it came to AI. The sheer amount of research and papers put out by them on AI should have proven to people they know what they're doing.

And then openAI throw caution into the wind and essentially make google and others panic knee jerk because there's really money to be made, and now everyone seems to be throwing caution into the wind and pushing it into the mainstream before society is ready.

All in the name of shareholders.

[-] Toes@ani.social 18 points 1 year ago

I think it's not enough, disable all the safe guards and let people decide if the output is what they want, hate being treated like a child trying to buy a M rated game.

[-] xor 71 points 1 year ago

But this isn't an M rated game, it's a transformative new technology with potentially horrifying consequences to misuse

[-] PsychedSy@sh.itjust.works 9 points 1 year ago

By answering questions? We are general intelligences that can answer questions. Oh shit oh fuck what am I doing talking.

Hey guess what, we general intelligences are capable of terrible things.

[-] PsychedSy@sh.itjust.works 5 points 1 year ago

That was exactly my point. AI isn't, though. All it can really do is respond.

[-] kyle@lemm.ee 11 points 1 year ago

I'm sure the military is so excited about AI because of its ability to "respond".

[-] PsychedSy@sh.itjust.works 4 points 1 year ago

If you can get chatgpt to drive your murder drone I'd be very impressed. Telsa can't figure it out in 2d.

[-] kyle@lemm.ee 6 points 1 year ago

My brother in Christ, the AI revolution is more than chatgpt.

[-] PsychedSy@sh.itjust.works 1 points 1 year ago

Call when it gets dangerous enough to be an issue.

[-] DriftinGrifter 1 points 1 year ago* (last edited 1 year ago)

Im bot sure if you are aware but thats litterally what makes ai so usefull its just responding to external inputs and doesent habe to be programmed value for value because it getsTrained with datasets and chat GPT isn't gonna hurt a fly the reason its m Rated is because the idiots who made it didn't filter the input content whilst web scraping its litterally too stupid to Funktionass a weapon except for misinformation which it outputs regardless oft its age rating openai is just a buncha cucs who switched to a close source system and can't actually make any gold company decisions

TL;DR: Fuck openai

[-] Socsa@sh.itjust.works 1 points 1 year ago* (last edited 1 year ago)

FWIW I work in the field and agree with this. LLMs in the current state are not so dangerous they can't be released to public. Generative image and video models are a much bigger threat, but that was largely something which came from open source.

If we really want to pearl clutch, it is NVIDIA which is really propping open this Pandora's box in terms of putting the capability in irresponsible hands

[-] PsychedSy@sh.itjust.works 2 points 1 year ago

Are the cards really powerful enough for so much fuss?

[-] Socsa@sh.itjust.works 1 points 1 year ago* (last edited 1 year ago)

Definitely. An A100 system is around $10k which is expensive, but definitely in reach, and you need two of them to run a 70B parameter model. Possibly one if you are clever about it.

And you can still do plenty of damage with a $1000 consumer grade GPU. Most deepfake videos are trained on these platforms.

[-] xor 18 points 1 year ago

Okay, so let's do a thought experiment, and take off all the safeguards.

Oops, you made:

  • a bomb design generator
  • an involuntary pornography generator
  • a CP generator

Saying "don't misuse it" isn't enough to stop people misusing it

And that's just with chatgpt - AI isn't just a question and answer machine - I suggest you read about "the paperclip maximiser" as a very good example of how misalignment of general purpose AI can go horribly wrong

[-] elbarto777@lemmy.world 9 points 1 year ago

I was going to say that a well-determined individual would find this information regardless. But the difference here is that it being so easily accessible would increase the risks of someone doing something reaaaally stupid by a factor of 100. Yikes.

[-] PsychedSy@sh.itjust.works 5 points 1 year ago
[-] elbarto777@lemmy.world 2 points 1 year ago

For you or many others, for sure it won't be complicated. The world is vast, and the environment you are in is very specific to you. Many other kids may have phones, sure, but they are not in the same environment as you or me.

Some non-sciency kid will have a hard time getting to do what their edgy mind wants them to do, unless an AI guides them mini-step by mini-step.

[-] PsychedSy@sh.itjust.works 3 points 1 year ago

I don't think AI, especially chat bots, will be more useful than a youtuber. It's not particularly easy to make powerful explosives, and gun powder is kind of trash for bombs. I'd imagine chatgpt would blow up more curious kids than aid them lol

[-] PsychedSy@sh.itjust.works 6 points 1 year ago

I mean half that is deviant art and you can look up how to make explosives on youtube chem channels or in books. It's not hard to rig up a custom detonator if you can get the energetics.

[-] Socsa@sh.itjust.works 2 points 1 year ago

What are we going to do with these Universities, just teaching anyone how to be a chemical engineer!!?

[-] guacupado@lemmy.world 1 points 1 year ago

Jesus some of you guys completely miss the point.

[-] PersnickityPenguin@lemm.ee 4 points 1 year ago

The last 2 already exist it's called stable diffusion. And for awhile Bing did it too.

[-] Socsa@sh.itjust.works 2 points 1 year ago

ChatGPT was very far from the first publically available generative AI. It didn't even do images at first.

Also, there are plenty of YouTube channels which show you how to make all sorts of extremely dangerous explosives already.

[-] xor 1 points 1 year ago

But the concern isn't which was the first generative ai - their "idea" was that AIs - of all types, including generalised - should just be released as-is, with no further safeguards.

That doesn't consider that OpenAI doesn't only develop text generation AIs. Generalised AI can do horrifying things, even just by accidental misconfiguration (see the paperclip optimiser example).

But even a GANN like chatGPT can be coerced to generate non-text data with the right prompting.

Even in that example, one can't just dig up those sorts of videos without, at minimum, leaving a trail. But an unresticted pretrained model can be distributed and run locally, and used without trace to generate any content whatsoever that it's capable of generating.

And with a generalised AI, the only constraint to the prompt "kill everybody except me" becomes available compute.

load more comments (25 replies)
load more comments (26 replies)
this post was submitted on 06 Dec 2023
264 points (100.0% liked)

Technology

60042 readers
2349 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 2 years ago
MODERATORS