91

Over the past few years, the evolution of AI-driven tools like GitHub’s Copilot and other large language models (LLMs) has promised to revolutionise programming. By leveraging deep learning, these tools can generate code, suggest solutions, and even troubleshoot issues in real-time, saving developers hours of work. While these tools have obvious benefits in terms of productivity, there’s a growing concern that they may also have unintended consequences on the quality and skillset of programmers.

you are viewing a single comment's thread
view the rest of the comments
[-] beeng@discuss.tchncs.de 3 points 1 month ago* (last edited 1 month ago)

You're missing the point. If the program doesn't do what it's meant to its YOU that didn't use the tools between you and metal, correctly. LLM involved or not, it's how you've described it, in whatever 'language' you chose (natural or Rust)

[-] firelizzard@programming.dev 2 points 1 month ago

The key difference is that compilers don’t fuck up, outside of the very rare compiler bug. LLMs do fuck up, quite often.

this post was submitted on 11 Sep 2024
91 points (100.0% liked)

Programming

17352 readers
213 users here now

Welcome to the main community in programming.dev! Feel free to post anything relating to programming here!

Cross posting is strongly encouraged in the instance. If you feel your post or another person's post makes sense in another community cross post into it.

Hope you enjoy the instance!

Rules

Rules

  • Follow the programming.dev instance rules
  • Keep content related to programming in some way
  • If you're posting long videos try to add in some form of tldr for those who don't want to watch videos

Wormhole

Follow the wormhole through a path of communities !webdev@programming.dev



founded 1 year ago
MODERATORS