Counter article: https://jadarma.github.io/blog/posts/2024/11/does-github-copilot-improve-code-quality-heres-how-we-lie-with-statistics/ about the original statistics article from Github this talk and blog post is about: https://github.blog/news-insights/research/does-github-copilot-improve-code-quality-heres-what-the-data-says/
If you rather like a reactionary video commentary to the article from The Primeagen: https://youtu.be/IxYN7DKefmI or watch on Invidious, a privacy focused web YouTube client without using YouTube directly: https://inv.nadeko.net/watch?v=IxYN7DKefmI
I’m neither a professional programmer nor a user of Ai but…
Do you think your experience, I’m guessing a pre-ai trained programmer, is reflective of post-ai trained programmers?
Will the inevitable reliance on AI in learning and training, will creativity of new programmers drop? Is that even a problem?
I am a professional programmer and a user of AI.
With current AI, it’s going to have absolutely no effect on “creativity of new programmers”. I would say it would even help with that since it greatly lowers the barrier to entry for programming. One of the things it is actually quite good at is explaining basic concepts, which can often be hard to google.
The thing it isn’t good at - yet - is writing complete programs. Especially if they aren’t very common domains like CRUD or parsers. So you still need to know how to program.
At the moment it’s kind of like you’ve got a friend who has read a ton of stuff but isn’t very clever or reliable. Amazing for finding things, looking things up, doing grunt auto-complete work, etc. But you can’t ask them to write an SPI driver for a radio module or whatever.
Maybe in future they’ll get to the point where they can reliably do the kinds of complex tasks that most professional programmers do, but I think that will take a while (and probably be close to AGI by that point).
In its current state I don’t think learning programming with the help of AI will be much different. For that AI makes too many mistakes. You have to check its grunt work output (which is still faster than writing it yourself). You also cannot trust its explanations, because it hallucinates too often. AI can nudge you in the right direction and can mostly help you, but often enough it can’t and you’ll have to research yourself. @FizzyOrange@programming.dev’s ‘idiot friend’ metaphor fits very well.
He obviously hasn’t. This is one of those things where some people feel threatened by something, haven’t used it, and feel like they can comment based on how they imagine it is.
Reminds me of when the iPhone came out. You had all sorts of nonsense criticism of it from people that had clearly never even touched one.
I have, but in my experience any personal gains are lost if I account for the extra time needed to review other devs’ PRs. The volume of sloc submitted has gone way up, but everything runs and looks fine, so the bugs that do sneak in are really nasty little things.
The experience of using it to fill out, like a wall of config changes, or a bunch of repetitive test cases is good though.
My team tested it out for our company (17k employees) and it was so bad we immediately said no. It wasn’t just harmful, it was actively intrusive. I’d be trying to type something and it would autocomplete the exact opposite of what I wanted to type. I was constantly deleting what it wrote because it was nowhere in the vicinity of being correct. The same experience was had across everyone else that tried it.
Claude on the other hand is wonderful.
trying to type something and it would autocomplete
I don’t know what program you used, exactly, but if the autocompletion comes automatically, you’re doing it wrong. You start typing, check what Copilot is suggesting, ant if it is right, you accept it.