cross-posted from: https://sh.itjust.works/post/25239919
Oh wow. The article says basically that but without the /s and then it gets even better. This is according to Mister AI Professor Ethan Mollick From The University Of Warthon and the link goes to a tweet (the highest form of academia) saying:
The problem with calling “prompt engineering” a form of programming is that it isn’t like what we call coding
In fact, coders are often bad at prompting because AI doesn’t do things consistently or work like code. The best prompters I know can’t code at all. They “teach” the AI.
Which is just great considering the next excuse in the text is:
this is due to insufficient reviews, either because the company has not implemented robust code quality and code-review practices, or because developers are scrutinising AI-written code less than they would scrutinise their own code
So who the fuck even reviews the prompt engineers’ code sludge, Mister AI Professor Of Twitter?
Whole text is such a sad cope.
Soon they will try to fix this problem by having 2 forms of LLM do team coding. The surprised Pikachu faces will be something
developers are scrutinising AI-written code less than they would scrutinise their own code
Wait, is this how Those People claim that Copilot actually “improved their productivity”? They just don’t fucking read what the machine output?
I was always like “how can Copilot make me code faster if all it does is give me bad code to review which takes more than just writing it” and the answer is “what do you mean review”???
Wait, is this how Those People claim that Copilot actually “improved their productivity”? They just don’t fucking read what the machine output?
Yes, that’s exactly what it is. That and boilerplate, but it probably makes all kinds of errors that they don’t noticed, because the build didn’t fail.