555
you are viewing a single comment's thread
view the rest of the comments
[-] sheogorath@lemmy.world 26 points 11 months ago

After it gave me to implement X in your app, you need to take into account A, B, and C 5 times in a row when I asked it on what steps needed to be taken, I cancelled my sub immediately.

ChatGPT is getting fucking lazy.

[-] webghost0101@sopuli.xyz 14 points 11 months ago

Much depends on how you prompt. I have definitely encounters this but some basic default instructions around it can tweak it for any usecase.

When i prep a gpt i give it a list of information and steps it is supposed to follow and the further context about the job and my system. Stuff it can assume and what it should never assume.

Creating a good gpt and prompt are alot like programming itself, i can definitely see how if you just want to get some work done that having to do that feels like wasting time but you once a gpt is working its ready to be used any time.

[-] UnRelatedBurner@sh.itjust.works 6 points 11 months ago

every AI is getting bad. I'm using some off-brand (some might say indie) ones, and they're slower dumber and have more payed options by the day.

ChatGPT if not for cheating in school I wouldn't use it at all probly. Phind.com was my go-to favorite, now can't search the web for shit gpt option is payed and it's just dumb. You.com my friend used this, straight ignored my request yesterday.

do other's have the same experience? If yes, follow-up question are they dumbing down AI, and keeping the power to themselves? 2nd follow-up how can I selfhost a good AI and what do I need for it to work?

[-] WaxedWookie@lemmy.world 4 points 11 months ago

Look into ollama and mixtral variants... You'll be limited by things like GPU memory though.

[-] miss_brainfarts 3 points 11 months ago

Mixtral 8x7B seems fairly capable to me so far. It's just that I need to wait a few minutes for it to reply, given I'm running it on a 1st gen Ryzen...

Any models you'd reccommend that fit into 4GB of VRAM?

I've tried Deepseek Coder, and it certainly works well for quickly churning out bash scripts for whatever purpose I can possibly think of

[-] WaxedWookie@lemmy.world 2 points 11 months ago

Similar story here with an old 10xx GPU. I've just started tinkering with dolphin-mixtral-8x7b, but it's early days.

[-] miss_brainfarts 3 points 11 months ago

Dolphin is just peak amusement to me. I've asked so many weird things of it that it's basically just cowering in a corner now, afraid of answering any question for fear of getting another kitten killed.

[-] UnRelatedBurner@sh.itjust.works 1 points 11 months ago

what'd you say a good GPU is for it?

[-] WaxedWookie@lemmy.world 2 points 11 months ago

An nvidia card with as much memory as possible - the newer the better.

I'm also beginning to look into the dedicated accelerators like the coral, but at first wash, it looks like lack of onboard memory will be a massive bottleneck.

[-] UnRelatedBurner@sh.itjust.works 1 points 11 months ago

dang, I got an amd one. Should I even try? (16GB)

[-] WaxedWookie@lemmy.world 2 points 11 months ago

That I don't have any experience with - I hear it's harder, but not impossible.

[-] centof@lemm.ee 1 points 11 months ago

Perplexity and poe are my go to for this.

[-] UnRelatedBurner@sh.itjust.works 1 points 11 months ago

I'll try them tho, already a + for perplexity, I don't need to log in

[-] centof@lemm.ee 2 points 11 months ago

Yep that one my current fav.

this post was submitted on 25 Jan 2024
555 points (100.0% liked)

Greentext

4636 readers
1605 users here now

This is a place to share greentexts and witness the confounding life of Anon. If you're new to the Greentext community, think of it as a sort of zoo with Anon as the main attraction.

Be warned:

If you find yourself getting angry (or god forbid, agreeing) with something Anon has said, you might be doing it wrong.

founded 1 year ago
MODERATORS