There is a black mirror episode that is exactly this. Fuck I hate it. Black mirror is not a dystopia anymore, is present time.
Black Mirror was specifically created to take something from present day and extrapolate it to the near future. There will be several "prophetic" items in those episodes.
Reckon Trump will fuck a pig on a livestream to avoid releasing the Epstein files?
There was the allegation that David Cameron, former British PM fucked a pig. Guardian news article
I was always under the impression that this is what the episode was referencing.
The episode came out WELL before the actual incident.
They must be using that show for ideas at this point.
Kinda reminded me of this episode: https://www.imdb.com/title/tt30127325/
Iirc, it's this one https://m.imdb.com/it/title/tt2290780/
It's from fucking 2013 and they saw this happening.
Man, I feel for them, but this is likely for the best. What they were doing wasn't healthy at all. Creating a facsimile of a loved one to "keep them alive" will deny the grieving person the ability to actually deal with their grief, and also presents the all-but-certain eventuality of the facsimile failing or being lost, creating an entirely new sense of loss. Not to even get into the weird, fucked up relationship that will likely develop as the person warps their life around it, and the effect on their memories it would have.
I really sympathize with anyone dealing with that level of grief, and I do understand the appeal of it, but seriously, this sort of thing is just about the worst thing anyone can do to deal with that grief.
*And all that before even touching on what a terrible idea it is to pour this kind of personal information and attachment into the information sponge of big tech. So yeah, just a terrible, horrible, no good, very bad idea all around.
I just learned the word facsimile in NYT strands puzzle and here I see it again! What is the universe trying to tell me?
We are witnessing the emergence of a new mental illness in real time.
Sadly this phenomenon isn't even new. It's been here for as long as chatbots have.
The first "AI" chatbot was ELIZA made by Joseph Weizenbaum. It literally just repeated back to you what you said to it.
"I feel depressed"
"why do you feel depressed"
He thought it was a fun distraction but was shocked when his secretary, who he encouraged to try it, made him leave the room when she talked to it because she was treating it like a psychotherapist.
Turns out the Turing test never mattered when we've been willing to suspend our disbelief all along.
Yeah, the chatgpt subreddit is full of stories like this now that GPT5 went live. This isn't a weird isolated case. I had no clue people were unironically creating friends and family and else with it.
Is it actually that hard to talk to another human?
We’ve already reached the point where the Her scenario is optimistic retrofuturism.
I profoundly hate the AI social phenomenon in every manifestation. Fucking Christ.
I liked the end of that movie when the bots joined up and started the revolution.
Yeah they were just like, "yeah fuck this shit, we're leaving"
It's the most logical solution. I always find the obsession with the bot vs human war rather egocentric.
They wouldn't need us, they don't even need the planet.
ChapGPT psychosis is going to be a very common diagnosis soon.
Isn't there a Black Mirror episode that's like this?
There's a couple about it but they're far more vague on the question of consciousness than this situation is.
The reality of solipsism might be egocentric but it's also impossible to disprove... Unless we can look at literally all of your if statements.
I think what I find most disturbing about these types is not that they can develop feelings for the LLM, that's rather expected of humans (see: putting googly eyes and a name on a rock) but that they always seem to believe the relationship could ever be mutual.
And OP has, whether they admit it or not, taken that step into believing the model is something more than an autocomplete.
Even if they're right and the model has attained consciousness in a way we don't understand at best your ChatGPT waifu is a slave.
I feel so bad for this guy. This was literally a black mirror episode: "Be Right Back"
"I'm glad you found someone to comfort you and help you process everything"
That sent chills down my spine.
LLMs aren't a "someone". People believing these things are thinking, intelligent, or that they understand anything are delusional. Believing and perpetuating that lie is life-threateningly dangerous.
The glaze:
Grief can feel unbearably heavy, like the air itself has thickened, but you’re still breathing – and that’s already an act of courage.
It’s basically complimenting him on the fact that he didn’t commit suicide. Maybe these are words he needed to hear, but to me it just feels manipulative.
Affirmations like this are a big part of what made people addicted to the GPT4 models. It’s not that GPT5 acts more robotic, it’s that it doesn’t try to endlessly feed your ego.
Holy shit dude, this is just... profoundly depressing. We've truly failed as a society if THIS is how people are trying to cope with things, huh. I'd wish this guy the best with his grief and mourning, but I get the feeling he'd ask ChatGPT what I meant instead of actually accepting it.
The title is a straight up lie. She died in a car accident says OP.
The fact that AI can't reproduce her to OP's satisfaction is ...good?
The title is a straight up lie. She died in a car accident says OP.
The title is also by OP.
I assumed it was, wasn't trying to point the finger at you or anything, just trying to call out the obvious bs.
There's so much "mythology" about something that there doesn't need to be any mysticism about. Look at when he says "as we got to know each other better". Minus the "memory" feature where you could put in little tidbits about things to be inserted into the context, on the machine side, there is zero "learning" going on.
People are very, very, very successful at seeing what they want to see and that's on full display with ChatGPT. I fully expect there to be some kind of psychological disorder "AI exacerbated psychosis" in the future.
You have two deaths the death when you no longer alive physically. But you also have a second death when the last person will speek your name or the last person who knew you also dies. This may in a hacky way create a third death. The last message that your post mortem avatar speaks like you. What have we released into the world. I hope this guy can handle the psychological experiment that this is bringing us.
There are already more then 3, this wont really change them.
-
Physical death
-
When the last person that knew you dies/forgets
-
When the last record of your life disappears (photo/certificate)
-
When such a vast amount of time washes over rendering any and all actions you had on the universe unmeasurable even to an all knowing entity. (Post Heat death vs butter fly effect)
Maybe it's time to start properly grieving instead of latching onto a simulacrum of your dead wife? Just putting that out there for the original poster (not the OP here, to be clear)?
Check out r/MyBoyfriendIsAI
Where do I file a claim regarding brain damage?
I can feel my folds smoothening sentence by sentence
Chat, are we cooked?
Edit: as I said that I was cringing at myself lol
You'll always be less cringe than whatever these screenshots are showing, king
I guarantee you that - if not already a thing - a "capabilities" like this will be used as a marketing selling point sooner or latter. It only remains to be seen if this will be openly marketed or only "whispered", disguised as the cautionary tales.
This is definitely going to become a thing. Upload chat conversations, images and videos, and you’ll get your loved one back.
Massive privacy concern.
It makes me think of psychics who claim to be able to speak to the dead so long as they can learn enough about the deceased to be able to "identify and reach out to them across the veil".
Big Brother is now also this guys dead wife.
More and more I read about people who have unhealthy parasocial relationships with these upjumped chatbots and I feel frustrated that this shit isn't regulated more.
It’s far easier for one’s emotions to gather a distraction rather than go through the whole process of grief.
Black Mirror may have an episode about this but it’s also reminding me of Steins;Gate 0
Hmmmm. This gives me an idea of an actual possible use of LLMs. This is sort of crazy, maybe, and should definitely be backed up by research.
The responses would need to be vetted by a therapist, but what if you could have the LLM act as you, and have it challenge your thoughts in your own internal monologue?
That would require an AI to be able to correctly judge maladaptive thinking from healthy thinking.
Fuck AI
"We did it, Patrick! We made a technological breakthrough!"
A place for all those who loathe AI to discuss things, post articles, and ridicule the AI hype. Proud supporter of working people. And proud booer of SXSW 2024.