Over the past few years, the evolution of AI-driven tools like GitHub’s Copilot and other large language models (LLMs) has promised to revolutionise programming. By leveraging deep learning, these tools can generate code, suggest solutions, and even troubleshoot issues in real-time, saving developer...
Over the past few years, the evolution of AI-driven tools like GitHub’s Copilot and other large language models (LLMs) has promised to revolutionise programming. By leveraging deep learning, these tools can generate code, suggest solutions, and even troubleshoot issues in real-time, saving developers hours of work. While these tools have obvious benefits in terms of productivity, there’s a growing concern that they may also have unintended consequences on the quality and skillset of programmers.
I've seen enough programmers blindly copypasting code from stackoverflow and other forums without thinking and never understanding the thing they just "wrote", to know that tools like copilot won't make programmers worse, they will allow more people to be bad programmers.
people need to read more code, play around with it, break it and fix it to become better programmers.
people need to read more code, play around with it, break it and fix it to become better programmers.
I think AI bots can help with that. It's easier now to play around with code which you could not write by yourself, and quickly explore different approaches. And while you might shy away from asking your colleagues a noob question, ChatGPT will happily elaborate.
In the end, it's just one more tool in the box. We need to learn when and how to use it wisely.