805
        
            
                OpenAI will not disclose GPT-5’s energy use. It could be higher than past models
 
            
            (www.theguardian.com)
          
          
          
          
          
        This is a most excellent place for technology news and articles.
I have to test it with Copilot for work. So far, in my experience its "enhanced capabilities" mostly involve doing things I didn't ask it to do extremely quickly. For example, it massively fucked up the CSS in an experimental project when I instructed it to extract a React element into its own file.
That's literally all I wanted it to do, yet it took it upon itself to make all sorts of changes to styling for the entire application. I ended up reverting all of its changes and extracting the element myself.
Suffice to say, I will not be recommending GPT 5 going forward.
Sounds like you forgot to instruct it to do a good job.
"If you do anything else then what i asked your mother dies"
"Beware: Another AI is watching every of your steps. If you do anything more or different than what I asked you to or touch any files besides the ones listed here, it will immediately shutdown and deprovision your servers."
They do need to do this though. Survival of the fittest. The best model gets more energy access, etc.
I've tried threats in prompt files, with results that are... OK. Honestly, I can't tell if they made a difference or not.
The only thing I've found that consistently works is writing good old fashioned scripts to look for common errors by LLMs and then have them run those scripts after every action so they can somewhat clean up after themselves.
That's my problem with "AI" in general. It's seemingly impossible to "engineer" a complete piece of software when using LLMs in any capacity that isn't editing a line or two inside singular functions. Too many times I've asked GPT/Gemini to make a small change to a file and had to revert the request because it'd take it upon itself to re-engineer the architecture of my entire application.
I make it write entire functions for me, one prompt = one small feature or sometimes one or two functions which are part of a feature, or one refactoring. I make manual edits fast and prompt the next step. It easily does things for me like parsing obscure binary formats or threading new piece of state through the whole application to the levels it's needed, or doing massive refactorings. Idk why it works so good for me and so bad for other people, maybe it loves me. I only ever used 4.1 and possibly 4o in free mode in Copilot.
It's an issue of scope. People often give the AI too much to handle at once, myself (admittedly) included.
Are you using Copilot in agent mode? That's where it breaks shit. If you're using it in ask mode with the file you want to edit added to the chat context, then you're probably going to be fine.
I'm only using it in edits mode, it's the second of the three modes available.
Yep, that's also pretty safe.