cross-posted from: https://sh.itjust.works/post/25239919
have they tried writing better prompts? my lived experience says that because it works for me, it should work as long as you write good prompts. prompts prompts prompts. I am very smart. /s
Oh wow. The article says basically that but without the /s and then it gets even better. This is according to Mister AI Professor Ethan Mollick From The University Of Warthon and the link goes to a tweet (the highest form of academia) saying:
The problem with calling “prompt engineering” a form of programming is that it isn’t like what we call coding
In fact, coders are often bad at prompting because AI doesn’t do things consistently or work like code. The best prompters I know can’t code at all. They “teach” the AI.
Which is just great considering the next excuse in the text is:
this is due to insufficient reviews, either because the company has not implemented robust code quality and code-review practices, or because developers are scrutinising AI-written code less than they would scrutinise their own code
So who the fuck even reviews the prompt engineers’ code sludge, Mister AI Professor Of Twitter?
Whole text is such a sad cope.
developers are scrutinising AI-written code less than they would scrutinise their own code
Wait, is this how Those People claim that Copilot actually “improved their productivity”? They just don’t fucking read what the machine output?
I was always like “how can Copilot make me code faster if all it does is give me bad code to review which takes more than just writing it” and the answer is “what do you mean review”???
Wait, is this how Those People claim that Copilot actually “improved their productivity”? They just don’t fucking read what the machine output?
Yes, that’s exactly what it is. That and boilerplate, but it probably makes all kinds of errors that they don’t noticed, because the build didn’t fail.
Soon they will try to fix this problem by having 2 forms of LLM do team coding. The surprised Pikachu faces will be something
Someone will have the “brilliant” idea to fix this by having chatbots review code in 5… 4… 3…
I could swear I’ve seen a shartup with this pitch
will try check tomorrow, rn I’m enjoying the sounds of the first thunderstorm of the season
Thanks now you’ve sent me down the rabbit hole since I searched for this and clicked on the first ad: coderabbit.ai
One of the code reviews they feature on their homepage involves poor CodeRabbit misspelling a variable name, and then suggesting the exact opposite code of what would be correct for a “null check” (Suggesting if (object.field) return;
when it should have suggested if (!object.field) return;
or something like that).
You’d think AI companies would have wised up by this point and gone through all their pre-recorded demos with a fine comb so that marks users at least make it past the homepage, but I guess not.
Aside: It’s not really accurate to describe if (object.field)
as a null check in JS since other things like empty strings will fail the check, but maybe CodeRabbit is just an adorable baby JS reviewer!
Aside: the example was in a .jsx file. Does that stand for JavaScript XML? because oh lord that sounds cursed
You’d think AI companies would have wised up by this point and gone through all their pre-recorded demos with a fine comb so that
marksusers at least make it past the homepage, but I guess not.
The target group for their pitch probably isn’t people who have a solid grasp of coding, I’d bet quite the opposite.
JSX is JavaScript, but you can also just put HTML in it (with bonus syntax for embedding more JS expressions inside) and it can get transpiled into function calls, which means it’ll result in an object structure representing the HTML you wrote. It’s used so that you can write a component as a function that returns HTML with properties already computed in and any special properties, like event listeners, passed as function references contained in the structure.
"When asked about buggy AI [code], a common refrain is ‘it is not my code,’ meaning they feel less accountable because they didn’t write it.”
Strong they cut all my deadlines in half and gave me an OpenAI API key, so fuck it energy.
He stressed that this is not from want of care on the developer’s part but rather a lack of interest in “copy-editing code” on top of quality control processes being unprepared for the speed of AI adoption.
You don’t say.
For some reason when I read this I am reminded of our “highly efficient rail” which often derails
LLMs will save us from having to work on features now that we nearly ironed out all the issues introduced by Kubernetes.